var/home/core/zuul-output/0000755000175000017500000000000015135005530014522 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015135016332015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000317072615135016247020272 0ustar corecoretikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD 0R~ƋI_翪|mvşo#oVݏKf+ovpZj!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\[.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €'} S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|;/zE'`9ƭZr(&e'd,LFlPh ۬r7%׿kb8_b|r wFuRI%T۩Ѭza\_/2vw>- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہwħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?~=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%J6TIsbFʶ褢sFUC)(k-C"TQ[;4j39_WiZSس:$3w}o$[4x:bl=pd9YfAMpIrv̡}XI{B%ZԎuHvhd`Η|ʣ)-iaE';_j{(8xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Z;#es=oi_)qb㼃{buU?zT u]68 QeC Hl @R SFZuU&uRz[2(A1ZK(O5dc}QQufCdX($0j(HX_$GZaPo|P5q @3ǟ6 mR!c/24مQNֆ^n,hU֝cfT :):[gCa?\&IpW$8!+Uph*/ o/{")qq҈78݇hA sTB*F$6 2C` |ɧJ~iM cO;m#NV?d?TCg5otޔC1s`u.EkB6ga׬9J2&vV,./ӐoQJ*Dw*^sCeyWtɖ9F.[-cʚmD (QMW`zP~n"U'8%kEq*Lr;TY *BCCpJhxUpܺDoGdlaQ&8#v| (~~yZ-VW"T- 0@4ޙ-did˥]5]5᪩QJlyIPEQZȰ<'$VO"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))?' c9*%WyΈ W\Of[=߰+ednU$YD',jߎW&7DXǜߍG`DbE#0Y4&|޻xѷ\;_Z^sнM\&+1gWo'Y;l>V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~SJ^{vn 9 j1шk'L"cE=K]A(oQ۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܫ1z.x62%z].`Gn&*7bn) u{Hk|v;tCl2m s]-$zQpɡr~]Si!ڣZmʢ鉗phw j8\c4>0` R?da,ȍ/ءfQ 2ؐfc}l 2窾ۉ1k;A@z>T+DE 6Хm<쉶K`'#NC5CL]5ݶI5XK.N)Q!>zt?zpPC ¶.vBTcm"Bsp rjﺧK]0/k<'dzM2dk–flE]_vE P / څZg`9r| 5W;`.4&XkĴp 6l0Cз5O[{B-bC\/`m(9A< f`mPіpNЦXn6g5m 7aTcTA,} q:|CBp_uFȆx6ڮܷnZ8dsMS^HэUlq 8\C[n膗:68DkM\7"Ǻzfbx]ۮC=1ÓOv$sY6eX%]Y{⦁# &SlM'iMJ았 t% ~@1c@K?k^rEXws zz.8`hiPܮbC7~n b?`CtjT6l>X+,Qb5ȳp`FMeXÅ0+!86{V5y8 M`_Uw ȗkU]a[.D}"\I5/1o٩|U戻,6t錳"EFk:ZM/!ݛ@pRu Iヵvyne 0=HH3n@.>C@{GP 9::3(6e™nvOσ =?6ͪ)Bppًu_w/m/0}T>CUX\!xl=ZVM\aٟ6h㗶E۶{O#X26.Fٱq1M k'JE%"2.*""]8yܑ4> >X1 smD) ̙TީXfnOFg㧤[Lo)[fLPBRB+x7{{? ףro_nն-2n6 Ym^]IL'M+;U t>x]U5g B(, qA9r;$IN&CM(F+ hGI~Q<웰[, qnriY]3_P${,<\V}7T g6Zapto}PhS/b&X0$Ba{a`W%ATevoYFF"4En.O8ϵq\FOXƀf qbTLhlw?8p@{]oOtsϑ`94t1!F PI;i`ޮMLX7sTGr!oapckL>2QQWo/ݻ<̍8)r`F!Woc0Xq0 R' eQ&Aѣzvw=e&".awfShWjÅD0JkBh]s9Ą|ק_;%X6Q@d 8&a)a.#ۿD> vfA{$g ăyd) SK?ɧ#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%˧Byd}gs9QNʟ. /ӦxbHHAni5(~p>/O0vEWZ nY3 cU $O,iLacoW1/W=-kqb>&IL6i}^^XpCŋ݃k-$pxbڲ&6*9mg>{rtD)wQ`pkKyt1?[ˋZ5NhfӛŮ Qu8Y4?W֫/&W˸~%pqq{% ?K~,#/0'NZ׽Kq^ėSJ6#j8GO[ PCbʍN^XS&}E9OZ]'t$=tnn&nu [}Ab4 +OLuU{0fIb { O݂9x 8/{&Ά+4*Iqt~L4Ykja?BHyݪkIf-8>V#ہll/ؽnA(ȱbAj>C9O n6HNe">0]8@*0)QsUN8t^N+mXU q2EDö0^R) hCt{d}ܜFnԴ.2w⠪R/r| w,?VMqܙ7;qpUۚ5Tnj ۝jlN$q:w$U>tL)NC*<` `)ĉJآS2 z]gQ)Bی:D`W&jDk\7XD&?Y\9ȢG:${1`+i n8=%Ml%İȖb7AޗuV3A7ำqE*\qb'YpuHƩҬV nm=Ɂ-2=|5ʹ zi ' ׹U>8bK0%V\ t!Lku`+]c0h&)IVC)p| QUA:]XL/2La[Xѓ F;/-rtx-rei0hE˝ݸDt#{I} `v;jUvK S x1Q2XU&6k&lE"} Q\E)+u>.,SzbQ!g:l0r5aI`"Ǒm O\B!,ZDbjKM%q%Em(>Hm 2z=Eh^&hBk X%t>g:Y #)#vǷOV't d1 =_SEp+%L1OUaY쎹aZNnDZ6fV{r&ȑ|X!|i*FJT+gj׾,$'qg%HWc\4@'@—>9V*E :lw)e6;KK{s`>3X: P/%d1ؑHͦ4;W\hx锎vgqcU!}xF^jc5?7Ua,X nʬ^Cv'A$ƝKA`d;_/EZ~'*"ȜH*Duƽ˳bKg^raͭ̍*tPu*9bJ_ ;3It+v;3O'CX}k:U{⧘pvzz0V Y3'Dco\:^dnJF7a)AH v_§gbȩ<+S%EasUNfB7™:%GY \LXg3۾4\.?}f kj· dM[CaVۿ$XD'QǛU>UݸoRR?x^TE.1߬VwխmLaF݄",Uy%ífz,/o/Z^]ݖF\\UR7򱺹...m/~q[ /7n!7xB[)9nI [GۿsH\ow!>66}եl?|i [%۾s& Z&el-ɬeb.E)բA l1O,dE>-KjLOgeΏe|Bf".ax)֒t0E)J\8ʁ,Gulʂ+lh)6tqd!eó5d ¢ku|M"kP-&ђ5h ^pN0[|B>+q"/[ڲ&6!%<@fpѻKQ31pxFP>TU?!$VQ`Rc1wM "U8V15> =҆#xɮ}U`۸ہt=|X!~Pu(UeS@%Nb:.SZ1d!~\<}LY aBRJ@ѥuȑz.# 3tl7 ]وb Xnݔ[TN1|ttc‡-5=VrPhE0Ǐ}Wd|\aD;(;Ha.]1-{s1`HbKV$n}Z+sz'ʀ*E%N3o2c06JZW?V g>ed\)g.C]pj|4逜*@ nBID f"!!*7kS4޷V+8弔*A19`RI/Hй qPq3TY'퀜+/Ĥ'cp2\1: 0mtH,.7>\hSؗ΀ѩ آSNEYdEcaLF&"FhQ|![gIK v~,Jc%+8[dI368fp*CDrc3k.2WM:UbX[cO;R`RA]d+w!e rr솜[/V`+@;Τ`5d0ϕ_Lع`C"cK>JG.}Ε00e>& 2䯫vNj31c$ i '2Sn-51Y}rE~b>|Ď6Oj~ebIapul9| 3QtUqSCxTD7U9/nq.JYCtuc nrCtVDƖϧ;INOKx%'t+sFUJq:ǫf!NRT1D(3.8Q;І?O+JL0SU%jfˬ1lމZ|VA/.ȍȱh M-r ~[0AG꠭y*8D*-Rz_z{/S[*"꫒?`a;N6uilLn<Yllmb rY״͆jqTI!j.Pٱh s!:W_´KxA|Hk1nE6=W|$O -{]1Ak$ ѫQ6Plp;3F$RveL l5`:~@c>q,7}VE-Q8W70up˳ A¦g/OEU:غA>?=CۣPqȅlW11/$f*0@б 2Dݘrt +qrx!8 J&[V =͋A,z`S,J|L/vrʑ=}IhM4fG(Ȋ1{TT%41Oa'$ 4NػmW4~ %/q68[&3@(5E0v R|ecJ;mǤp`nNXLӼG#"U`l7Ya%ݡw (-TSS,ҲPE)ۋRE#sY>K0tW p`M0OIw+p6ag-w4˧fyzq3PdlM$BL>:g4e,mR8:u%+ŸH$0?I.ZeaZ Q,˜UtVw Q4Ÿ7i< Ϲ>ݎާd I3Rheg>~ KvMȦN8i|rFRчR68g,j5R˗cďf)YsލH,c_>}`U}ri<轔pRoY:qb ,b?xQ{Xk0 p ͐B1I[hʳiU6EZ^/N4jN_m%iN %{1g9j3Z*Lrr,?}|w3Hj;V޸߆F&$l^d4FbA4sقht4N4J䙽a3bdG\VHwہa0=<0ںE<GF;urp*s#Dm2|Fciب[eyZ{.Rm갧[,vxоLh`[F% J[ʀb|UӁq+[K ҄iX VхOT4֡sz`Ά`{QE u=q4 ZCBrtٗ넎_,on8WS4BƄܜ]y4\z.n/F4'8K|Dj7ۣ#?u)!ꍵ@ۘ*X #r 3"rGƈ"BmOim' =̇@S՚hN 3\+-5ׇ8$kE5JVDOj5dud>p]Ew ]5oņчyF4)P?Syv`}jYam<ؐuxe-Psu%zҔU[ Ll]5Y}H~8:Z,=Za '%+oIUVu3ΩR nJa%˫}MG34 _"/>bj[ItR yuG 4`[NXu9Ջbu uK}]U (Q#L eGAcLm݅k`ۻl*q!D]6`Tw3U`6OӤL8¡]ۘ9Yvx< O)ae! 0 uYǃ薥x[ H||:ܺ.hG3lxuu7FC!2%[-46ftf3xng2#%SL^¥A`o&HC1x kzuЬ/ˣ{HZt '׵ap,0ZNZA!6UDD KɈ3hzUnUtn@B]Q2{UsYⴾ6JfMCrnhj,`FS}P+ af뱀G hPmx@VYrEx\ Qkۊ|䓦ZFҺB%\*1 ȼ:Dk4Um4eFbn:B`VA^=":27=6XhT_PRh#ySdâm[ 5,Ռ?jucO^}xR\@X]2hpm2nŪJ)Vd-?Z6 quUp_(e&woFJyksχ3S~XbIԡn%UJ32Hbob=cvsI97+8*pD@jkqI걏}xaʶftg.F2.;4 ۊ_Ռٞq<1fQ q$2#ۚa -'B78ND&k]EzԚr=x7%Bh>HKY^&ẍyغ}(tnʩiBo%Xp[ gk -WuqUDA@g7Y}⮵x|&严IdtKyxmʖkz?- 銈+6'9֜h*+u>~u.z щ,kR[p8}ǵ7mw; &y,zXj*>j +Xn~[l͈3|!Dy ,лdNS>Stɲ02g|U0Ãe<fzmh><KPҠu{A@WxGc/ wq-O{pSari]-`&Lܸ4wf2][9*gjU금N0!Esh(fb.ʛ*㙧.LS/htű;z#HpU)oje\o/R2&U,Bzϕ5WJoWWs\V- u'ZX¿{U,q*-Iu!L.ZKL$eS5+}mҸyqSaӛϛރg uY@"2Hd:;.X֎Dq_"r!tW R!W#2pIy;ғUqyOP$Dzty^8;ҷRDq9/rL|k:;!`z)RGYDi "YZ Q"YB>JcuMnKvv= r^=EyHç$g?>C]M$w hT"£#.?<;3![KQa[/HG;ֳ I9N4E*EsǨ?L]LG+e$f0[Iۻԁ +}PG*ۮ nM)ܧܧrj]NTxuֲwLAIcFGhBhf+%+ŸnshkP mʘ44VꏍdU2gjq{o2C_Qݺ0 D-gz-lkanGŪՃ |S1\$HDudn Ysv&|_7^ASzxC45X:㚧zZmb#A%,=~"b,sNlRET\ U'GxǛy ˥&YS׵~a?V{G<{/In\ee#jJ8ZzN*ͽ g),$<]ΑpdjX#:ĕ*wY5.3TkEȗ5ȞuEdmnt`YoI X% "o?=س|88}S8@|h 1>@y](7i7<üeR!1} ay8pt*bn݃p:0ֈOB\k}&#23.} ,} 1ݷ.(b*O[TI@ae~dcn9 mIE @1aE2h;Iv=1|gaN*K a#l bh h-!w$P,!{1uv)E^%b/NV J.spCQ'x+`K}q*iфXޕtvH ]w#x^N u*aE"Smr;g '#dy w X\[Ndi4b,) FV]_,,׻t/ C=pig֡ ;Z^] b508e@]bagݡAMHhn0&P ѽ5witf?Nې;]! : N^iy^GX0m2Çv́gx=G۔u-mshп"u;i&6 HLIfGL(n0= ?ϕΥpJP9Isy9}c C{$ˣ\Sn02͕wBW <NCb g18RTrGBd#83Զ0NL gtv4ଵOKsgp[< hܡK a/}$b;AC;F|x*wL-}4*@v8J^/7=ykyoD2SGJp 92kQާwxC:uziQ<ЅjLrh)REP55G7E LM2 miG8G4˲D8XX"&X$r,IZ +Tf ,Wqyc1ԕֹ*aw"!+i2RrO 00aq,^= ~<%Z#JA`lA3|_ś}e۬(xYd6G>V¥YNNV>}D̏?pAjohuXZ8¨J PFQJzs8@ Fg h6)쓟`yf"-Y#P E2\ɺ )7eݣi1lv#.feg= __Y #&A}jq)Vۓ9["ijYũ#r\T ɺ\,U4~m,dՑިz| -J?~?W0Xpl7Gv?T?2sґ\6;DORkqVKvi%/edJOƠP1XL㊌`#"o4*q rr w{1z1`)ŭKz ¦0CC!NQ[j6 p M];ilbr/ҠHlmef勄G"ћVm;rpcLnj7N~fiQ0đ"dE򦂵m"/Qt*vx`|#lY`iQ`"%t,؎;FO7B8j>Et+aZ{a>ķؼ~uq\\# 3Ev^[)nA(2j bPy$;|wOi'(;a£6]7r1y$I,LRxlg,cen?N8A#ѦUn"~F[pF="t~pÐ6%?ǕyPO]Hx6gU,R.oꀼ{gkb\q,:9jU5v -4d #V:&sה*`x.Ũf].k ~`2t9Lp3:yrnUa!>;Q<sígt*%d=6=h2K`Fbɣ&Qܨg_q) q Vp7(LȀ)k0AP n /0#S \{] K}zZx*E {A܍``6 `؜l ]n};qPmm> o/.nXNzQFso3W>ZvNC`{۪ȡyl~dw|&ymi꭭VrV|o pz\,.& >r)7EP="U1v6K0~Ѐe *cX_ut]PtwA ([ (]P4A#e; u) >MP;;w?MPHAzz;.4AG .OAWr 3t*c[XPYq=#" Sխ_2ߣ j`ZCyg9l 'WyU!\b Eф!ܓJvkJ2N~&=0۽BTI=ȓw>-? Ix98aw^~Oy5 0/*#0>؋Bw)!>l as sQuޔ6lFe<ƓXjp op͡edp=/@g峓,TaLLB5Eƒ`d+==%sςQxx $K(h6|֪P;;30$XO;k% A*իG?_A\5Qр H%{e:RSHzOstfb oNEwT,,gM9WKYuu}ig4 's)]vtL̒ pZ wmar^/Nj˽\8/Q6t.ͰjtP]H{'U?>1$~?I[3 eb0 ^l|֗K|Q/ OV!lTqvMPvcZ=qCcF"nf!(GݜF 5Y`T^`ȧB"68=0&1q0 Vx֋A;'HW sWJi(jB/!nVJ i  s`@tQCU1:L@T'bH*LM.a#ؽp,n$Q̎X2O`""qEhtGm ;C'a'8IO3 fo SW vs^?LpuChjܳε3ce&GHU6q=HI c4Ktov4ƚJ)g7Һ}/&0}6i O/16 @NX1G7`iZ5c6_YcTW=+)~tﻃ2.**\4N 裵puexyLN&Kh4"h[yԚ%PTz|ŸSu&`+#MI05©aZ&•HR-󽠀1p]_ #`L; 7} 9Z1$*2Ft+eQ-6m6ډV`,)!LcA{x*r#8a"]eY=&[ r,i<78-Bc 4Vuż̆kߌP]XsWSAs\.eBg r,){=`ߣ7¡;6P)3-/ dye6 rpz(bpjBtz^%mD7L8:}q;.B+읊R"ۃb:/f{ၢ7i_A/xà.nE'naOzB,9zzYgU-5GI/\uQNRhr 1iշa nuJNM+3_̦o]^l [[ ծGGi_ߩw5ds 24< Kx#a(C(JP dP qdj/b3!GU<4;&(SF_`Am<͡[7_|?gV{";`\miV`dqIITFh]몛Q g6^N&nGD3NUAH~F=gk`4Nd[!y׶FdeežyhQ$%[_(ֈ ;_dKBUGR^&?u:9Rqdp.isK&=ôv\y|4f́MWLԆRJVC#esn6xm&dt5B;'܁?xQӆ"yBO:7mJOIBv7Уڳ9i^Ҙu0e^Ĕ6nwI/I,~;>~z.:z菿C 7C +4z:RUks*fpb$ &q"ꐕE0}`:eɼEST䖳M?\^^4=w 8'p (.3]Bd)Fϒ<x'F *0; N֙>p[*C?VIbXupRLHYeF~Ge Faxr٫[$8Nj#p /Pl\jd-uϖ5oY4q8b$]n$8%Oiף6.*<2\M1cb:[ÒŠˆjG$H }H]#BUZUL+X9c-G))OU'2o'ȱʧOB Xx-1%"Oa3Fe) A{RyO~ +pCJZV-H%GDន>U i2~5z4 )`/[hL&#;>G%6ktly S lA1F!^Wj|CbUp+%Dl5&ۂT!"I78=^Hpr0e+)&5b ɻ/EzMƈa}Cb'㹾;t(G8 넑JN MV͞G;Vk++ ,DxaYʠ^kznN:x ϏA&yOyﯢJ˝Q0bNA{/;m4 AӦ Z1:S, F͑޼rp!\_֚ Ť{ٺAÓʯŰA!L޾J^Mq&,I<dG@'Lwa$D&{`q O\5$А`gIvg^L(FH͈t켆G'[Ctj\hHft/%UâT #F -QI(gb1gzc$ jp iwDo/uVn&-2AHbCd>OJ|$䃲zQ]tB.`#")j<oKZދٴ5s$ })zpWqT*²Puap k` p>`7mPM0=J xjMB^j߾;>S]A[~{ 2fH`h5LDNhsaQmz=# 4SyW$!vQ Y?fHw[ qZC560ZLwCk,$<|Ti1D\}>^5IR8n[Wy se<>&b>*zTYVF**}_CLx` (<'' 76m;aQòJ%%$=VGo.p&e}{Ŕ5̓T/]6$V(z7 DٌP`Zhi 51xBydHpL_~ۼciOݝ [NMZ0guޯJ%9npjnzQU¼77i$4N' 9q0Zr}Xgc. }ƈ;ȓ6N/zkh3jSt X̢+p(psPsIxH{CFs̝z-q?_I҇^EuxMڄ1^q WA?ԸȚզ Ӯ>pjB qDH1iq7cZހ5BefԤ؅L" 2F<6O&<{͆d;$>?b,)p .inFB*I|ur8#^6Ƅ"Jv`$NgoHpm]s,K"Zw25!]I8GsR\R kqa)"Xo$8::罓pf_LI޿y{S6\!q(c-]!ZL!F (w lBiE3i`<gʉ(/sN1p #5ϙ(2?#8#8/3Z$VX^132t"a;$'Ŷ@w@@wdRZB2VTL!F " C>/*ϛI<2aߣr4kXdB̀ G>7}!q0kܸjMp vziF!~hN.Teyv[L#Π |;)co?_s$8ߍ'Wt>Z>tZߐ JVg/¢6*O:41fj}lbPf^BP^ٱ6jE iut(kF( 4ibĐj0&Mo4\zlQyHb=x 1:ӣZĭeǂ\ƪR[C~m%jVLYҵ<56hUsi|Q>/z5J6cIŗZDxT$Jh_rQEިW*mIOYXSM] 0nB6˹[/2v]C3>wef}.އ1{7ٞR/$‡⎮[ 6d ʎgxyZ0#Dp:[ISpaTiiUbQ &iEsΰ?xp\ׅ3E(/k7'=' 8rnpIHÌfq.jnLl#ԇ'8JDB{Y\kH` |Q[K&"fkF҄c29PXEyO/s%KbBUM1>ҲUwm^ְ$klO\/$v0gmbtEzV8G?.^VvS{ǃ E(5%G^=O.LYH~Җ_ZjgCt' //qeJ_K[{Xm_c|M;gc+SȠ^shPp"t&xeަ~O7믦P[6NeZRɻRǖ~-i`Ѱ alZD l71)&LY??2= \J,D,pA$K1I|L6)`) ./;1yך': b#G6j]QBD3 p.](pG `"H>z?%!7̞n]ko9+ ;;7~08;71LJ` #KޖĻ~ljrh궝GxX:U,_{[@lkz"LK8($_lvoR0M͌RKl/3;-1A*'xyt[eCw&L5t3"(23{q| LWGLIR(䇓¨HwϓGѦt+k!rT\XWnx(ߍN? ƴĭi i9fh.W!<*kQgLh,}{}B[}DmǙ/lc4|dk/eC̠pCH2sY9*u_W׻U28DehI$qVL G1hpy{ՑJq_Z1h:^,Zz8>}O.2pDawI2P0UfotFW[N-s[u[N݆/ѰVvIљ&~:QCzY0`+\..g~.;z`^#޹LƽH€#GPc41>h'|<er|hZXN/s0Oʨ:H#"ӳ{D&a#wF+@*b@c~^MQlol&b;̜*aBQ%ҋG'a,Y:-l-}Wxbv'''G'ރ] !Ar2(Ȓ78#1VӢ,P.6^9ڇNxAʈX+-g)xKj[ٱ^i|޸̙)HjԏT/c3W?p[c'Q%"7[*3lwMեn .L9y3{Yś}RTܭ?vW)8' ٬9/'83s/T%YnĜ)3]}u)fT-h%w2 "/`- K+7 %bTLf]zJvep73,XʥT(R .K3㙉XD]+V2lFRHO0#NO;Y1@L O]A30qS8T[榘Uq}8-r.s),⟫?= ugAIt}or}SfY8-ʛEnM,^ 7i%-PZ*t\OD0#CH$bB3/-QΗǛDcP.԰uunV ťhP\ %/4AiA T,2Z6Pq1ͷ3;i6TZ S# „I+jȯp xQG &=Aă0[cAp06lê#ZhU* V-╂p&kk6ڤ kBJ(e?BzKhUkSmhX'l= js*ƥnQ 7w-Iƒ [/XlgK<>𱔥ئ~ҎRW(2֗[޻nj8/Xj9NN0*L`QibKѹ]^BSFĴ\)|vK`E?\4۽mJF!j4|T=:ٷ_=xwqW<k,׈I s/E\*N<w] nSoOcwx:n%XyDuD`-N28uAgҺ?JD?ۍ T;6ei :աɈЏ7*Q! ]ݳ ]~ps-)>'6;W5Wd12l)mݙ| α@;gtDa[Aw|% p. b#RqÒ14fW~Uyw/fR=ͳxߺbW- 9q$*[9ϵ)rAQP&h(cJ| @mW+[SjԌ2x2 %deQJ)8UYB"Q{ϐbq3$W.n/DJxEl`;go\_K*;ͧ8]]~1]Gf_N?ƹLzqD06ʸι9gҡz)7C 96? N 5PNR%&w]3T5F3e FIDqiw\ùJ0` QsAluu]3 #1.`38T G#.w,4VK N:@]6M ͨP=Q5Cjc3֥D1ƹ΢p},q {hG h#-u^d{Y4#3.5>H'+o`_ =}7oX77u8|  Uj5Vf"V&g βdP PZz>}"y2ǭmgBaFL#ix=0?NO96>YZr]~t[2p~u;-ysK&e9|aVp^`X]U?<&%5}FW~چRk'}Ce߿fWpmcǀ')p<AL#I輙Jzr݋53VM zUʯeOgE=N~97nwp U#Rh1&\11oS_i ~,_Ⱦ=͏IT!jȽ==Y[sՙ9~4\ZehFZaq;Rke1) *g1ҟ>{PChtu @3wXE@[٫NL'_`w_uN4˵QvR 9|uS 5>;EJ(;K[m#3, Zk$6`єb+},0L@9X98*&X\:fnF-@)Efʍ=ux׺b%eY"'sPxŕ: =Mȉ$?`$ITN+~x[J.ZѰڠRx}wMP1uyjɐ-0c*Vʨ[bHV Nl6p"Z PuAK+˭./ *w:rVZ[#.i-$jf D-NlXy עUƳR[4Oz:@JT=Bjs \PHJo?m TV*긐uK.>池(vȐÉ yj<@pY 6ǹsh?.DRYhhIFZ81Xjnָn#Z8mrcX7T"ZDE/~E;Ҙ Q.fi@؟ |$jP*k/c">ۉ[؀ FP X>PEm8.*߉ʱxM*]HEY |.nnu]-* ^sKO9xw)Ԍ>\`ޣXqS^|nHPH 3fi(# Af!p4W2e.rYFOje5Bh$V Ysp% .X<XD,&09m\1YH|0~s;9>0N\Kⶑ >Dv+m u؛@<A/b[p%;UWD4 NFmӊ۶xnm^ʳ.FZ0n x,W8^3po)`G`h 'sTpecC g0SK-`,5ΝPP` \Hr ԘqnpR3KIMa4D=4SI=XDXik'YSQ+ !LK $ +I" 3ӂeN'[Qgrrzľ'ПO~mc0G -kB&<8% 6l` W}>K0S& ʈsMsε: FuFB qJh\}FB;`aN䛓"]h6vwx8B76nkLmiYYQQ69[F=A'|j6 %9+D1A͡$e,T2t*إKԾ̥Vi)URiݡljM&RE#P^ KTPUMsA\(a%dTNk+Tp>’4!H$HpL+clY0$=Җ|*br+BF(NM-G$FI(VA,CeH8Y$4JűZSiXE H<~! D PFԌLB9P,ݐ#40z^`DpN`9iа%Iv2jِpj]vz#iҎTqk6M,% 87 B0 yYTRɕ`K(Pdzi{nD& 4"ƍkudt|Ո`Bf4ʘJ4g$ }l'fkݐZڡA>]K9H+Z1ph 5Y i77# 5" 4B_oD miF.粦8i0% sf O8g&&_T8A,m TaW.f^|_+Mӄ^>zޕ'֥.;1+a6XLiGh1w)1/+*;tI,`~@$/lcآ*4ky{uޗ LaM_Ϡ3gG~QІGYf!!zㆦ̐[)hѺPǯvug2 N*Zb&VL8u.mF`NO>rف, XAD}ٓ獕mH=}39;qCrn!CRf\m Ցr5"Sy{dҀڭB'fo4$X$#tN.6췀~*ofKha "Al;7q.`z?x8H~ý wx\(d',qfyuxݼDKo p\-"= Mz}Wg˳6]@in~ "L>C]9?6W AXk~B;ZP%ks>VD|[`v)b{co1S}cZdC*Q-Ȳt z<`^]| .z7@/j;OnAL\__(+w< SV0{eG$+P"Vös 6L0J`4oewg\]@] ?'UCt{$r3ڊHݸ~ 0U_v:S97y`UH̏3~vc j=dtbUZťüѽ2ǃe_1mˆMuϜq4|a;i)VjWU=SZVɔN&%jg W{O|q% rUKl)m:#sf4ٷa|OvPNXw-gÈB7I 2Q#(CEpE3# MJƧKVeA^I8n˩Yv:90\S93kkGn5 QlK8pwݍRKFvM\K#>G;/FngYI4X wF JVϾG})6լ;mzV6W-ۭG~k):7w)HCmZ0'6 );fo&: ^ ߂{ L Oߌxc`E 1Fwko`&-c[יub{*a)]L6wˤ=0R⎒ƍw y,Ҏ_\s鉭kr&dZOq‡nȣtx>8aa8aa)14 6TGoo-eP;reA/Ⱌ_|4B9S62/Bm@Ҭ"M$aà3AgRS AD81"lmx~3fRzzs7/E7f.?&LQy֦#tœKNg8Fk"w^=hU>;`N n(cvu->?5ҰYX´V;+Xl ZKc`}%:t=_jRN./os_(01Ce|O>H 5/..')P~ !~z}Z`E#x.rz#Z5<ѻ7!>y u?7:Jף>.C]ַ*Z|)I|1ƿN>닏m$LQD7xyE8H^x'`.S{mˠ[ x!^6K}al!BjiPtq}l>6˥z4qzrB0yg_P^B&66Dhs;LwI{U 7)W+|`SouzMd"|PTah{Ue$WEGPSj]ΠehUP,{Y< ;zٷ0 FdBUJv,q}#e+TIq{9kb@׻+;< F.LmS~"/eQ)z/J!JއEy8m+lWS8ێQz93VIT:7e0[mK{\g=e]fпDxMHZ!Ӛ8QM`l@B*N82&NX821""*4t4#Ջ)92$&5䱜3g0)Kܷio+`8r~%;WٙK2)pc؁gP,xWŞ0*Q֗ui}ُagJX) ]Wϊr^zK Ajf9&JVN?cu=dAvf7f eGxx;T֗܅9,^\Vϫ!fUϣ q¿WS0,KP_ 6/ST!}tumwV"9`{1IK UQ]By_gsEt|O!|PmӃ0m'm9A~umwzpOncZ:vjn('wr\qWo=*j|{ 69<>@H0)ڣ֘h3'Lm`A;yB=9{9 05[Yx lLHMc-HD: 5zdvwCӽm*M>B0mGǤo5LEZx8jaD;z0i2`ngkڔ4045&Ք60%IH%+b`;"㶖 2Q̄B)ZK( HԡVnF*D&&HZydrh୆Q[ݦe[>ydsak'J9h?|෾ :Tx/bZ?I"YJEDxLSXF#-BH(&aH0(6**zI5m2H@&A&LEq$9?y">pauJlQ֯w@h%륣%2UYyJ^g**M;-Hp+,σ4A4$yL-` \F2ڔ35:99MdLB+Lti4[m3 sMW0&r*n9mgsܻU'n&;%4#kM>|:J0MbJKJ2IWG?_r \ M'N"qRM[o;l&&C~r D~ Z$D؂r=vƧ^t:||ơ.ǡ,խy T 7tg"+ 8>%͠RIJX% F0%HV:ʴuQg:$!?AbNwVt8e`vִkINBeM9^z."ۙ4Rv- @6^@4v Ɓr-Hlul.&~6l.\೹gsB(;bM pM-%\R Z=ьrNOMϡN~O}ě 3S&Xj~HsJp<l) ^d>OMmv`44]xũ>}kM5/ӿvjkoG)X~?!?oo^>%ox?ET؉Fb]?M˦ژd{[|응 L vWc{S5pm.b}nn l|+V_F(I.Bi%BJ֛b)c/yzY׿ĕ`~}t[`f4kЂQCcmi!%IEJO:x hA[>r&|-!w\" M X=^EF 9F r @RoE6 /^d+w}. #ZGdXl?VĨy;Pk}C+G@_Joa/R̓l!RqI6cADžf-qˇh0qN/Z2%RDT+i%L]̣-tqkxn&$(Xk,\(3\ّ.1DZE\wY(axͲ*3Jb|(xhJA^S B6*ɞr C!7Ө$kH +H= c&?YApl| {]2,H"&%mp0$$<A}PB 7U 1n3e7^)?Y!wj:s?kH!H5%)-KJT֖㤃q;#h3jPxyaaAɥ#!)G\oSAo:ǘCAL-&xފB$t40 KOYm [̓-lqˇak(өap(òGv*_֛w~~Am78(qe#2m)F<ӨCE͛)oyۀㄨEJ\B&H5&H;Z> &k\ kxsEN b) Ԓ31O`- cI_yQV )l3M\ ,Ӿ|(֍mՖBQ' 1:\c.I͋"fތZ>mF͸8HEZsT%HȈ/3W\| jP$6w`%IQ[" "*$sZDDbj[>Taj}PL r-M1uḤ-tqˇ۲UӍ,G:hIi(%(n^I 2D.?nd퍌Z>jjb,QgHdroґ8xGKȘG9OCu Alɣ|JaK Fh[بC!EmzetLĴUc?|Fb_#$ow{kT2#& KR) (aI䲔0&JfkP1NQ*TR YғYFA5x *vjPP儩lsc8ar/~{Ҵp+1I ~,Io,&mG'>| @v4p,=M^:/(4]Ռ% (=蹯@<|(BW_S\LDpM"bܺB4t{/Yu>3B4q?6=b%%*%,5kcjNȹk<@v| wOC8WfI; X )D`gLy=LmjPȅ7dBf˻զQͯt;{?Z Q.t?Գ"+;x|e1=rUʸbI^VKТcVk̈́Q_gci gxj:p>쵱)Q`#,wgcҚďaA jP?(c[cyϪfdU!K6ψ4?ëCrn43Ƹ;O8@т|(U+uLXrAmJ@\FCMu3y'Z[)W74nbM`\PyI*ӿPҿj4N̈l>TKJ-5;O}[Ϯ&GeTM ي+ ĻQ}g^|v1 u\͖26 lu= גv@,ހ5#51<I\| OW:hyޫH&0XH60A8㨮!h5򡐫zQ qLyyXhj 4XҌ>+]3-Eti`nGPx;UfӼĠfα*,^u5T=͚܈5eԆL'90|Hyt#i;QˇByuST/iM3vFsi+Fs7_SF:e񱎌W+aXEM'|)ƍd$QI8T=jnNs٢i*yw*k#ضSM+TJF9r<ނ[|7bUUjK+[6j]jO0.%72&{td P!haqQ򀁾GՐ:,Cs+’pk9P ݏK!D:x@61ˇBnLTea=@L+f=MLP=o_6|\}>SfWr.'ʡʨ(wcݖK]Q"Qybu,15ngvR ;9aQvGQPM C(e͑"![OGO?FӏZ>mWys5[* ¿W=.Ms:c[ Lڱeꬅg 8[o_X>o_c~5[?eGf~~+}M7zzS5Ԡ 1^1.f$~;n%^:'28g8^a;rFx"n߶CaSd}pET}yVJQ7fҳ`溇lv=bv0–tV3prrAiɳO0ˇִG[@_kGlJq}<#@;b{a\L;0n}=]v9)>-“@wF,hOf"{A]\Z>>nݚ15s\0稨)A 5rX㉍ɀ͌Qs͆Y#M SSgx)yEqtn hC!WY#Ko7[szOg=Ӷ` WT$ćNC-]Qvn3lccz;q]o{L<*VA1LN}S'{y!_Gt"'أ'GdO57W49%8nG~$2Y>f;]oG3zczv9/ucVOdOrǸcLQA_o*L vlŴ.}0KVKMZ$Ĭך 'L@2 ѶQˇiQs԰@`^ӕ,  Ld[4y4鎨C!7rxjpaݪ)hqHD'i!%IEJO83;Dڙ7BDhBh͈q^'XΠ֗`2JOSsIJg _mI>.l ].^mRnK8X#"5c{+R"9)$ dnSgZhhڷ HiwLĩbIFOQ{7O_+߀yn dǔMXX1DZ0:f*ިi*"2{ oy4 0FW!ȯ9Ҕan&;O1J"s t_+f5` RP knR.gu`0.~*QCx}1k1_#rDȭ0d=G|u>ȯ"jzY())Fؒ;f.?ugkfkMRIZd9'-4c1OcLJw7M/[I7;׃^j3nu531v66EFN}2g\<8֕zzgTlJktA_ӻ;I9I[LL )_\z5HѮ"RiVۓZ$c1GLJ&tVD<o0F׷"V7~v)CH-څ+z72,!0huRɔz3#ɩw>X0嶯Ncbgf|{0ȯbI IRS}aʝŝw6M4!*ڗ\""RIuuK.di=8uz8^$ df@eyxL 󷺽6Jg>ui+SU,A\n\/ENgCGMt+|j<&ӹ%'p2g@|<-G|;?(7tf:9Jhܙ3>Y c|W9i>+JAi,NÙIi3j*^ n+'8hG‰円>)Ռhh^#$R3`RűϪQMFhNnاmI}0v#L!*1AXDL-E9ˊWB[t~S,ݶf,wRJ_9KuG:<$e̥ʜ]rWvo;.GQ1yuu0y#d.(esbKR`EKcmt2P;)U΢_#z\z>˦9(*φuu|}1(Eu4V(^_&JI(SwΈ4DJSH1ދz>uYj1|/TuqOEk2k9M*(()E8w"é t4[: l]i׈ddgu.DŖu7l0SxꙧDR'YZX R3%Uixrά``\] gT< Bj2v14C{[wh,q ˰B)O ]X-Z7%L]B"$LМ7gf7I57+%ayᶃF4f2:k>yKT3w=8kqF@;V ݭ3fC-> !p FEٹEPBAbxAFl;H1ʾ32zliZ>G)A(B5(P18|nDe'x[P<bfzc( z !0ʶ6UxAFoOt)' wH\Nam1 0%= 0Q!ԄWN w/7^XSokC4Ƹr"d+dL*9)YT4uԯv:VsMm–"㻃0@~C8gV5qf $dH]c™KYwp/!CNG4~R:?~!zHpo.8eG2*9{xSIW=;eY]-aš]z7t?Ԏ]d33?hG_H⌌)Ʈ/Uqe?.OVoo|f\Y5|p5C$ՀCl9%sz~6-_o(߮~vCbF W=/l˳.̚d Q4+b.?ѯ oGX,="۲lrT.|8}+YoT"ě1oG {|mnj#{nߋ{MxP4Μh 9O}9c<8#)QId9MSO8i}|m32?ynj'w>f?"45e|n|]V?^l^8ۨfnŲR͗"* O(zi'siXw'yne~mٙНeк: gެNMWBn_зbGq@- jһs8Oוy-r/۬٣kǷ£^ZoQjT9oT3>ˬuˋYɃ}@Fg,?zm ym LnDۜq4/{sB^6p"ce7GO',U+BK#d+7_nʯ.@Vs~//!ihmsÒ~xΑJ"b"H iw7r 9c8+0DZ@dĎ@aOkc !CpO#@b)QP{d'XkHAR#rDȭ0d=CGhE+ijc1Ğ)0UeGRШL4]u$x)ϛVf־ύ7s@>C zBUHA>DYPwD9!+w%KS-<遾|b ˚zh>ZAR`.gzk#^wX{͆C毀M7NqgpGuU(qB%Z!m#tJ >Zk-~ bZOc5@!NRTK%T!mSg:D qn{اB@"r7g%`2[Yo^{X@6Ϟ\YS?1L ("Bfc$9ib)@v$Vˆbgs}#W9ȃ##*X Z8v`\هL HX>12s,:5Ck)v:0n!Gtx(4`0Haᒰ'f!(.@CHf15upu4Aԅl@Fۇ78x)YᝡcnDJD^(;N:)'NjJg]ծѶ.l[B>{̥} %=uF"2`)(PpctBOW싚:xZ4ƸcXp%LEQDX C㯟d;ͯ z1|6P+*gs_Vͨ1BڛѦᶩLS8ATI$H)IQ͜H0UUcs|VGmfu3 znb=$ G9*))M FOc =X,F&I.F)g$Fwڊ暨mxQ9݌aEt.\qoLP 8uapRcv XNp%WSwOc %cmN%2;_Pspg"e}¡J@'>̹ ~v̦<Ш2f ˩MD0Qj*iNSy[S}1`~wj{Wva[xTcQߋVݻl;'=Vv]= 7]1C8w]܉G&&8:T< %#_尢Ŷ[oh-VjnyQxJ"*3v~~M՟z(AWEq wM!PQշroX;zL]7ۺe͛Ȭgvdѩ/]˽~fu=yT k^Z\sBrE]bE]?)wͮsO'$ X[TʠpEJ-zZ "҈N@8X\X WbTNJ$q&;+T"V8`DŜY <KR%(7J̈́R9l.cWԻ36 B Ǯ12r,p6ȸ֣dI-jx~*rhgSE5r'VNo̬D|+G|_yֿٛOl:ˠYUoEiŽ_Giu7Ϊq(.LKhO끇{\0د:8fultſ%XSGv y\z=^+D}Wz%!fd1u^Pلu2v^6^>tKk6n8y(ã5uttn6z4lԹc|%;ǑSM-2<JӲ(={ a3x0;̘ IESgesN(7dd}ߗ|nfMЙ5:38til4V+ܖ*;tGU Th Awb9ltBh̲~cG$5נ*P^*BiI$c*G<SQ GZ$VJ݂j (hw~?֔f7ϟ(Vxww`Q6'ǘFt^/ur{~[O2<;`02('T:h%|{'`D (q` {k%κ܃[R|OŭzmT۲RkrsJO ?uud L ƜQʻV0J W1Zէ_GNMDʽ`)Nsj5}غ!)RP>MQJA'DߜZkևphLi_>zm}o;ŎNHD3~bL'dޞX.=FPF90WF;G)sdUxeΠ<f%䈑W?P IQ)-ӆ="^FYP(c A%(XIH)#I&5YLV}-AUtw)">ڲ^ԛNs*J=kM"Σ ;,hSs{KL#sT&U,+صN  !6Y砇Z<#$a)(mojfgǼ^5nAqc^уAIe><+kVzEE⡏n͕C10ۮw𻘉DA*JU]MVˬ4g^3;-=,~b"C/~rT pnu]G$b3ă^)˴Auz NLnt[08=,Bo2Q!ߺ}0p~F/F2V!%j+ĽUX"v9R7rx\]W(Lrd*{h>ݝzox1\t:Q<66HA}NᭊZ+5X's73*)M(QkxrdŗM+zfzJ2gd1Y/8K1Y((d]|dMco즎CF/<"ӈ! f%jl!Q&b䪘Th ?T9Pqɿe"_$BUaU1çr.`Ys VqB?z`7 4E9q1p!|I< 35j-U!M&XʼnJ*N8V0Y_,XU}rh2[N^s܊Sb8~xO=U˽0mN`zi X!_rr^=K0|ިj0:)ϼep;T{mmw=LV/ɨA1Zxk~x!<)/^x['~ +>͒&gR[;+m6M"yќ3o.h6Ù5y^ SƾCOe’3P:7yw nOb4zRwFvIP|I3Ibp9#QsSrr͑%ҌJsq<[Q/`g+68ySm U5ijL6 gT%HVPZa=slG_7g*5Hqc54  ̟ͰpkEbL@M 08t l+%aWcA9R}sMF~[,.\ !޺=_vF|2M%Bk3?H i- lA`p"5""8tl4ٟt dW5K0B%(~0_t鏡K3g).Vq㻷}Ѭa'Ik$769D™2lIQ(} I\F @p_}`%5-_qiַ%PW)2ڂʼIs^trcŤ_Tj]22ׯf#eL-udqu5:A~Ue6`M?6SINfp;ڦ aklآ^Sx-cfgVo^nniڮ;k쿤~:dX3@AlWT9Մ %yYC9&8nSo7J`US/aׂy [5}-6<%R8Ci|8) \,$blCBMka+KKY-1(5)A?R l3bJ]A'YtYvwuH8R`&9!SMZ )ig(MYTOC " iy#ATD vl*Sq]SG|tч2ZE<8S418dx*str`ԋK~5BC%ZуDqHzfrt;wŭE_ ah!sRݸl0*. ؚ'3[U|+uD,g$q_=ۓb!I^^!Z?8+iMqy٤/ ˏ>{&m -n6zqgfoĘ*>sWvk&tkH,pk3N9|̔zχ>|NG܍3;granAwcp=CEkEU>ܲß\QhЌG_\jpۏM;ɀ_O|Ʀ(ͤ~y?mi/nt ̸;#b0*w,r|cwZ.GqK…5嗪1מ$ lӝy;Rn :/}{&A kkCt85*MX)h`s:ͦZg"Q,)#&U{*f~|dgg5#x5yLe"exr7wk%aZT^n0zg?<\ɠ:CL1|nF \tw_jksr3\zִ]hp+'iX3*>~kΨfaqӯ̃Lyl|FC.?f哯E齦ʐ[L'ۻy1w?Щew㯡&OqqѾA]; ᠦ׍ 6;ɀ|0\>}aY2;1co@S ]vf"d![5W3`XJl<*_1N"Z\51-wZO"E ![G~=>ŗ| `,YN2&]d8"LJJd 1yU*ǖ0l,>2`E~E~'Q2,06$0N~-KI9<}Ie _t/6mvƳbѯt|;r(ӚY#]vAaL#Z,q CĔ(e8A Qbpe@M2Hzemz|$ȫO7r/ sL(Rm\䜧>F,r Ƹ%)J)4?B&4%%O]9;?vv#>snGN]:rB *܊~޶d|:0EgGrn2M Ź;xtbaJzcUoaz pO*Mk'ڃwSQrQax7_ UzLCJZۯS*1- ѐ$轛PP6Om? `گ9FGB'GSvhc,>t!X5g._D_ ެD᠊AIxǯac10fHb$\gߡv_zT~ _*ޕ$"mN+@60LiF-a Od,*"Q122(Ҟn y;;gy.C^fP^(((#'vXma{sE䪛Џ DsJ8I97inVT8&[1D-h[Am^ӉZ6Q w &wȇk8hFcpy38crPqݗWbր5zndT\NB!#QdT=k\K>+M˹IQkXdz&UB+i4G ~zO1/k~Fy6C4"uML^(s%y:C h Ӿ:~{ix:? *BDd`i8  i*$-z\WPx|7Gv>_g\yI14V- 1r h Nu?6GR6ڑjgɂCJR}W枛Vå_4hJ>7.ݱ\Ah؀o={X?"7:\LQ@4 DA" OPhFcphZ/sP=6y)h1Q!fZ-3?)wz_J"W(V"MC /JTw4SA3_5lcT}kޘrqL%FA"́cD+w4>%NiwT|Y㵚%ĉU 1@VX=S?F|)4ڂ#+8zlEGL}. ߮{\<ܢmwo1I ̌(Zz,u %V+a_ i  =G%1!B%"0ATICϒU9\6K!p1p[ %`cE&>&EIor1JEl8nlX#Qk{h(%qI$3kG "#r{GxsA`xbQ\@<Fl58Cɭ"Vi|Mp sKj^z᷁|:6!xPLB9a# hy98Jhc)®&X)9`Fcp˫ywAb=zcuQd. y89$1b8۞OGp>4E@шfEJ|T2 T_3"+_Tg TˬUSxH}KJ<$!9 x ZNXih7}|QjSH'L>kAZĜS4r4Q_oFcp Cr/Fxzc@*j,385,qQ&Eh@5VS+#ނ|pp"Dމ;;wv^Uixé|.Yv"E0*-A=S_8yNG;W_2~pG#4XQr% @ǒ`0^ tJT/4%Z 7Z~4X3c%;-^Ã|Ϥ'i&1_Yc>z[{ /OYcJ6=qX1=b!][JjYcR>&qw=ߙ4|9wbDnn h y{F̘02ImQ +%ay@OQ5~{Bә#2@Jۜtv?HaWb|pݰDv˔;PFȔ[4ITڛ0Hp|FEe@H)4D1<R& IRgɘEɘTDͲ(V5U7KI*dbnfgK.BYTd(Tȓx|>wcʬE(9yPM0셍/;dhsh&Q%%;]\tO+FFQǑxjv\#w@cX-S`xdXH8RDh]R~ɒ#;1NFj\h T,+&,T:@*'l&u)j,?mE2Lʷ4dRûS4fJ*EJ3p:[v5^9-$EcwJ#s8Zdl~(F,vD9[S|IQI'FW[(Gpp$(7fES&bQeXB^jnaR&vMSMNs&%\0suRlG(n,d >݅~LecFcpn&GIOQtꜴ8^H)iFcp7[Bky=,Y)g,0ʨ hFcp*-)Z6iDJڟO!=̓EK7Ea2[ԬmWaulXԄ0LjIƘH>E[aB XNk'GmP\Sd2H'=vQ6oNhFcpDv.Q7b$X}FcpjLY̐-:O+^}m05ZdiWŜ̎Sԯ|qӍPʅ:))M6K@$3 e/MPZ=ҝ,(~5@18|.LE1iTP%MPYԽ<wn);@18p)%m@ԙ̊퉂nGxJF! h o\3Uٕ6@18f7FcsPL)(g4^>q{6AO"9<̌gI|KTs߿۱!Qo 8y>B Y@4#ߌ'.h ^I:MFIAp,2` Q2Eus@t<5ןG旴Sgb ^= I-V rQƔ*wa;?Y1 "VE5&UBZE]nmt4m /S%.<jZzFcpTlD`Q^hQI!Z-3?Q!|pT(۬(IkMHFra FKs Ѡ5>9ڻI:%n F¹#ms(uD#DnFUw;>]1TICQML6t,ɥr3 }Eh4GeOynTL"*2uEm h4AQ5%84*T^w5[=}?i,10"QOzFkpLԜԜwFSpybWnXyuƨqB 40$n2 1iϵA.cW'BO(NH P!qnؘ| yx ̝OL5aםXh͹  e{DS'5n;sSH6ԒX =ɣg 7FB$fD7gFkpwa\aD$ hG}ŨNVJ7hu9)協F(ɇENQcS_~4b+k h JV~U:5! |R C@t<;YM>_no9JMJ x&Yg o 8C\`3D'!6Qs=^ ϳ~<(,oS%#U ,$ϕ{l'mF+BE;@qIO3Tnv rf}YK؟$Ź%*&N Yrn LH$#3( mT`$|."$Ă^/=LRyt4|>E?6fJ)M~n:N3㰽*D17V"gEdt7+|8؝Vxr}Å[A+ޝ{N`/f2N>דsO#oyžKhzG7#'j6[_3v8P.N{~$cvnι|޻8''Gg#6 ۥlDn3΃T;>詟҆N, tsEz(?l?߿~p]N/!xbƘ8S@~&{Pur]'Wz_)WN<'ۏ\9ݾ_dZx8_ds`w Ns.z{$ɐly3| r$iKn-ΰ_=xzv9  !Ŋ-0/.KRmT G7(^-; ڞ骩ԍ@5vSe3شG| *7n_g͓^Ucmu1ɦ^抩(ESo'~GB M_O}2ʓ 6q'aEo lhTOZ6_;~_g//_cOvz~K8 kg;!,_Bۥ0.HKǗC?($hy J$lK+0~`BKΛh$ށ,7WKEߢkX˚|]zme]N߿+:+D";h# 7Y_O&IUQ_#FY.iq00H px fp/sm%`Ck9jͧ/dLsIP~}I D0^| ߙ J@ h S or" Yɕ;Rk[ÚUlt╙xGNG'ttZCz[Kt=|oܙc@JOGnH(^zD(%()!fԨ$G'OfbS 8$X@AA_H) 0/Ê0޲{جUKheIO!M}榒MCCdf'ȇCk>9HPC|D4\ };=^|Vx`/,4"Xf tRJE%õimr8BB_.B77 o^/ɗ3.<"d4g9AʃVR[XYÄ(gPZcOӛɧ?$8dzuIp_>`ݙ 1OFA)*&~z J{l tBmk,* aX~R!~ye38SKɼ;RM^lj@U#Za)ͬJ!0B2?Oٝ 낼Yvta0XYgq:-0Xy\h\)8#vhWR¿U NnU؁]?.GvjlG-AV3eRbBKNU)c|=(PZiq'dt ]q &,&"4TczV@\~b'= Z]wbntTStmxށg. C~VExk(J 1 (Y-R2(K33eLG0Pekb&O@A9DŽҐQPy:gX! 3R6'Y&"D4Z̮Z]F{Lt6N: |Y]koj vy`T#)>ȏzJ=SZTZw`YTx;ˏ?LXHuzދ=Sc"fa!0m |a>0jkjkBvjkcXc5X5VcXc5X5VcXc5X5VcXc5XZNkFLk2> DxojkoUm&cS\R'DiY8t$<$нwg5L޻m(-(qzW8G*qucgU)M[FlߚJ*f s]c:ڦeol Ze9N1Ն;kW(Fm[;? ?iA~;[ʅP$%O5C{Ifek:,[x}o9xEV?[^.eѿK$#># Z\HF80 $+ (B"<+?ɯoO%*"@ 8 fO5O5nf ܬpmkód9Ƕ{$7%@\Pr\YrzCb}8И<)m^|% pFjahltxg VOEcjOV- HInniyP> iRJZw~vU|_ Jͷ- \q4UFP&Q`H^劵 شb-:vZ%BJz;ӆ>WQ\.X+0hE\%5Sf3fV`Rh1d NFYjMq)edd~ I_!&Jq]=S<:8Cǚ+9|fyPūXO>>I~Q|wFv!c&1D-$E۵űv.(P[a WvR)o/1aaƛR6cƺ iKЦJ:PImKILtG;¢U9h{!10+TK0B¶锕؃ ϲRi']$K/~vɶC1$D!r/aazbXH=_K^Ǜ&gC&::Ӵf{Ozmp6h3RZ%|eE_: ጼKnhe{(G, a~8XnšEf)}z9 Sk &VRn :ƌƞf1aO9,N1Tk%i"X&}}{6:]~ W1l^q~k_Z9٪&V7#CbT!;GbMAh#¬0C``s È`khXlO "a "G콕a!0m 2{EWDB z=AٔYil4OQ*e 4ʤ`)%>! ~zCvILdmlK,]UtKt+ /<¬Uel0=w}?wt+PI0U"LUbbX-Vy*RZ^ !TիL?x#xTsxT,[,GQa(##[HN:W>eirx-5fWT>"週W^c-C=b=Yz4=P!{7 ~:!xsYf!XS<׆m$T8xJSobL:$pav'lGDf)WK(@8F3meTF)p#9q4rD~l[V([ߖށn -?XG>0Mh-ip.4lVPEp64mo:aY/]lr~gӞ#DX2΍R5 33'7TKe9RԘ-{ t^u !YzF,s*%Rx"%eRYLfJ3nK0Lyvdl݁_ ƝW!^Z٬m璱qafIb4S0QdFN=g«:AZ:.CPGSāq(7H;8.=" .Dь xnF:FHuуpҚp^g}f-1ر2cJtl$%*,RL;x'c >^~a'j;Nǡgɥ:>[êIJ"IvE{a$Է 0  G#LQB J1=3lpۛ΀sz~J T"3IkK(.㾊[[ߞ aY wK82zYH&cSb^,~+Ϯk!)X%)? EقoAM@iV]-6\}Wԉ\ҵ+%ks'7f_'^^.V0ĵXŕeI[WhzVO PSO6x}OWMݐnB*u~8*Өh8]z՟5OzU$zm+e3=:>MlhzC) 6q'ҩoerCxײ|GO_}:<{WƑJlvvTTFqbccac PgD"5"X UwAuK*† [<^իWz{G_yP}{p}7-66MAO] X5Ke\I~4ϋiFU9_ҧkA[Wa?Ib4|Kb#[r!?"J[ 41%Qbg\^q:|u< uRlV}k"uː ݛ'rTJI~Y3$ !Fd0Ff#1pGWض lrؘvk~{J!f /3߿gVFdkiJiֹ lsϼ\HnCNG\Y"YY>P|d⒏{{k~ w:{HnKq-}8^;Ц7 ]Kn0IA7QlyҟNZOg`(Ugy,Ɲjv|>0uR6YH`zf#h;ՙWmkcrZD\ö꺿չuak]|&5xSq4?V4;5$~9mZ:6s)+kjxkgc`͞tҸui5Ǔ8|%p>)*!2PP@8e&w Y{7|}T!Y7p<2OC!KmyzawY*z6oW~b/M8䗿"] gF@4Y`ܹ6 haz@0h6*/o^c  𚽡jg-hAٿzI_ǹ~=5 ZM"6 LC*Ggbɴf Cqc=nj@"Vs̀K`|-][|>7EvELhm7{&s7 f7ۦ d{,( ;9 ٩fFLhޅe u&k?HR?}MR:U5ٟa9-\[5,d֤#jzힷ•ĮW d=꥝U_SͽG*M9W\@ ^mmkqv359f%qVPӳYVat2&ֱHe`H$,?|MB4Ű}(c?#yzq$d@=(E71y Q@:C |7%6MMnIN(Bph|5E:JM_|E$<Q8F}Fb@^;!Ԫ=߇B 4Ha kq$GF,vDP4=#A;t5x`B1,=(TK gSZar| %Ku;(b 3^ќz >JZϙcpJ:a cRKeP 9!BF-cz0ڄ4=߇B_yX)IAɁ=αP,FYneG] %<7qdRY ,suP|?)>&ؖln`x 2*Cn ELG,L=9 >ت"JA=߃B ґA.RL>Ӓ Ρ+} %ÁR+8' "X@8k NCxAWs,V@bdXaXվ#҇B V# C<l2 8lTx",v5T$+$lMdEZX B A "#iZRf#L=¹\fCǞCxM'MBPJ NC>4F X")1tusXP0kŕ+&<>J7`^ψ" d4W pT6I[ f gĒ;uSp^u}(O)^2B 90Qr- i*$;Cx(PMx"8FH8T|ۿ{r& dRq\ܞ31qTF,u %Vî>ؓ:u2kGQ0ryPT&I>JS.Bb8.8?( xq(CxAIgJWx8i)SG}BfH" >KJ%F*49b% 8wI>aC5.7r!{ ć`QzőI:"ErN C'ouԿKao=8|*5#]s7sS̏!G]ֆmib";K-%~X8KaB䤮N{$sUֆNVڄq1:muuS7#vŪMCR\$Jesuw>&z+ӌAD;s8_l0/J2F,`69ıA{1>=I8:N(7N z6&GY6`.y^%Y~թUY"⯖^ Kkp,7)rA4`Jsjsu]<;ZsQ/aj?]5\kyYw>?n__7McaJ,N$G7ef۵PU5N^҄}%sqci$6#lnFb7.,iCLk\o:_,=[9:7QQ)}MnJ$f9j}GYݒm.jڔfuMzQ 2!h2_1w^E}wG{/G?:xwˆMSSx70VvElR׬ez}eo|kth]_䪏>m t/ AN=|Kb#[r!?"M0bhy\J)ԭa+4_$*DÜ-f7E%wj^ Xi1IrH0V$^X^i`*u^bcڭ) ͚dG-̿"P9 d9SOoNK=2"[N;MTJax}BrP|d⒏{{k~ w:{Hn~Kq-}}U}JXjJ o:^`lI5]MPi@- *I21IGB8%=Җ"oTxIU07O㑿X?_57bş))qcv8iUsBC#ZieR?HBO1<3€w`c?1)q"$"%UUz#+32###'Xl"CQ#0rHRuipDE0>B;2(*n'atNCޣGX}Scˍ; 4xc=NE!4CY|D"Ӂ8Mc^K,‘Rs{q*3E? VXܬ07o 'NӨMf: egX̴Q*c d6]u%nSTkd=;VMOlSmC򸁚s r*ꮑ[.{xvX /]s\" ͑+w= 2V'`Aes6?.ӹYdf&S~c x sZd 5d(V8db  ,QR:0rJ%h<ƀ$5@yBnۃf>6SH0f W,*$U2屢7Dj#&@,N*K6T9(X4A$orz$ȈR%*:I$8xf Q3l@R%*EsS\IƆyTZi!eVma `Yib:bIERAnǣۄ;ѭB8"0*ԑK9 84@)m%GBe@fik. 9uPpR);$<&\NDD^#8  ,t=UP)y]qSi'L%g281؊T b`QKDxL@S53[#H1Pߞ4 ib.(nT?8Dc$DY[_R[~stNzjp3^qm^ўޘ<>܌MnF,Yy=IR1!RR%pXj0$܌5݌v{Q;aWVw5VFk|ƛq`kay98X$wfWNDEW-DM88…IANpB] Jb=DXɐ?0p]Ϋ^]oW2Sp;NKi w}B#U.&`g 6Rj3ʘM|YFN:(gk:xa5[k@p7%&UD^kx(,tG~AB g-ȿh`!t&۽[cY^g\J!D!e!x0k5f,`ZFL&Z ih:i_~0 'ש 0d?zOnDD 8QCvY8ҒPN;roD5q^_l#J E[^~ukK%w\}vb-ޘMysFd>ZR S\-&x#32x5sf"pa#v-aஊܵQZ)yD$ME]qd@]V42aD%$@tY$h) Ӂi]IՄ 190h+ȅ4*TP85u4H`&$M `BxXGm4a{." #$d8XG!1z ",iP%LR#1s[/b~=9[ߐ=yPwk+uq66"EtsZ:d* 2F o2[cRd^iGjd^9HysQ bG ^x:˹uܯZ* z;\/b7֠KWJKu[ӽT "V=5:-TC UhU TkhkSn}dG!rFAf G!YkL`D$  H NJPe5@$*,uHKdcAGXHLiRR0fDpCLDB, %0 n56d_֫[_?iyK,w^3\Ͳ~omr qWqv< NY_?$xdlۗmtYYqTN خe|7 Mc8|̾xM\A~ BGG"DYM0&xwUn m~{Z5Y@ȱ׬}LuWm_HBH @Z܅sc̱l94Y mI )sgZrAftg)Ll R( jU'a9p^?OzR]RQN[TnK.`i,m>a6PndM'oޯft6e}?C(/.e&#QHYu2LwZ D佖豉hj` i8\~O{V _sx9R5nRMùg[W< kma,`~ $^ N؆_2+UE$$ߪflɦ%5=Ff"UU}Ow* DOԞ*szbK*IZ/6mnF CQ?UXUlQ{Yu Nt|3N'85k[{F?|z!B(bғ!no Sհx'>W\C7Y৮7V enSƿDW+_U4{׵ڇ9!r>th DS9DiQmAWoF;PRB'Ԫ")L{i{$~m뾱Z}Qh [ qD6pCZݦ}ʙnznz ha*7%vҠoÓu&]>*hm\2OVZ9ъ sen>g)Vy#Z_T o:pأP2v7@'.sxJ}ŒWnӰnutdNP{3?E+?a..㐈Y#'0ϏFPGa?asQ3+,1JtMl 9_.*gEۗl]!fcغme$NLi4q>jc `6vǃ~dPx`L _5D6P`jZ%^/t/y31;eُ7OV&ވ~~Blק!VDSyf3wy {bFgA;YWSScAZL-@o]AyI0RA\1bV;=M7|pm>~SNC>Rη/D98z5I$J [$SψJEa1#s3+%X9($؀8)69I/mjQ.>V#@7 guhti!#7LimBm`c,(!@IB3u:)~Lmlvv+8H4irQvڳ)}nժT5OtqtlVW =^gb;/v]y pr) /luosv@jOntLޙ[@x ۠mNHd0wr:2z9vDi7=1=&>-L(a$+)ܙwlX$9)y@ 9IxIR~@ yLrQ24n${!lg ,ZEv*gk.Sr!yPAШ.%+gYo:OHr G  ܵao _$gYDCJ)#=DiB R-uJ;u$ 6I # !Ax ^,B3@*,WCۗ˔1cFHv (mNP@% #bA2BA.Hb?4{YKwok[ c_=H-8HWQ4'ϙ,=+5ыJc 4&q~#~ۢ5T0x# oDroBE5e dVzk!zd:BxlGD=Jo͂3Ods?VnaҺ帶vvEh1eyeRضxQ:҆JgׇL"S*ojRkL)DRS'ؠPJ!Hk8Gvd=a1bU[(QvY&>ߺYE 9 NsM~6mfo~}?:ĵ-2ye&ܾYMɆy>yvut;xz!ÕŗMɃ\MW'"{| S~VY'KݟZ;nV1ה9sy 0"*a>RO` )XF!םH戾rNסh 6QjSVjyiܮ3k.V76Bvwbtw6[egg&,To23Eo}/7;pL_eϤ͋yg=,OdGej&j\Z*SG_*շ\lVVQ D5  N'c(ucc;S  AN 젰 @n:HVٯdNqi&q[z5-oO [|?5>t4zԾev7g:4M LI r3E cf%܅PZyd$cD(aĨBAJbv Ƴ#Tu!7R;ѪNrrpma04h rf st{ˋ$F2Y WWA(^X(^֠2aj`֫ MEeRv'zBE.z#+iwrM3c>M XN [tK*l7h ${#򎖭1߰I%*ǩrd6/*u\D}=*(W5>nxYC-yQNx;Jt 2Yl~%脃rV=/~ܨSj\1?qLyq >fu}5a3{m3Igc_++=L;L~ys~(=DC/ d9bI.a4Ǒߣm"Bm#8uGHepbG4Ɛ)f+$ D 0 Μ`~oS]!?(ϝ]1u$ >ȯ-D3|~Ilɟ pG/*im^)!%zV V=xF:]Gp2QY ZQ v.C/D:%^Q*JzᄕC5PV<ϝY繓?;\Y*#^uċxqe2w0.5#)aR #RBef1h\܉ϝP}'D;3!f~tS'47; 2#eEUi`Fr #=`gb$"%*\}W&8_x96ǹ&*mQۄgzʍD:Rn#}|oަ5W5l#o#Xmmowquɜ5';kHI>2u,W4{7Qq./kfsҁ1X]rیd5ёq}g7cIip09sH[ :F *xm)ӤLe,:uk[f[JCUФPbx֞qR (/2HKt,aw17q$AzM\m '@5 D(؝T2IZ쭹ʶϴ̔N3n}jw+<:9Z2B1{wg%^pz]85ef̒5ъ6Rɵ/=aC' ҿ[_m>чQx0"(_}ݵqۍ> ZV[xteQ+Y` 6N]8 3&2vPٳa%{|'كn=I؁dO޶gn=MޔJ yES2)ӦƚO’_/;׽m 8J nÿ/h'"6#ՠ-Q3 {JTyo-sW\|eS?| %I!6/AX m$zoX#De,1LFEu zl*V9Φ5[lS1?Ui3w%j?Rb<rZDo22ivV .KN4/ɘdO9=kg|ǣIXss3N19cvZ3!cȣ/ѭl)uM ¬y" , |r#3gO+kY{9*&8N͒_=W{RQE;<*?-VwE.e ) fiP\W"P cXoEʼ=9^yvP^Mt:H]v~(?Av6?[҃;1 u/4tTDᔐuo^{Qm+ ÏÏՉjXy6=rZG(z/ghh&D*,Q((9Ŏ "@P =/Pv-5q 0xj2Һr"/7N OH_`?ze.w:{r|W[XvMK=,Ĺ@~<6@QK(b\! 1,! bRzUA(1ܠ#Дx@ h#ȤZ2$O'y8SI^8iT4|+W&ugH+Or Ӂ*F-@v!ڎp?Gj 0h%8b$QD%'rQ3;?Kv&TkS]6 @y!b^\q]RetRVsܓ(Ql쬎EK N(DzrG@b8̔QpA"m:1$hn wT]AJFkȁќdsk,,mfk Be$t@0{%%FIт)}7H.H-z_8/ C¡пp_8@7C¡&mˡпp ̡пp_8/ C¡пp_8/ C¡пp8q C¡п/|@D#E_ȹ(9=Six`,(3xn2,Ub2o%'6vlyeֲng``D`3$ RE-DkfFhb1]IQ)c*Cz _jRz\TkXeՎGNЎ5AߩI5mTmgds"J Pz .ym tUZie=KN8hGӎlB7+$@9 :%&  *GJ=,ZZXf߇.B?yU@ +2C]>hrBԽz*UkS{PTN BUjYc Py2ҍ$ˁ [>2f]Sroc| AHRVaDwZ D佖^MFS[)ุbo;h2&H EFV~8!,Ѝ$Ṡ{+%RDYxIΒ\5t pu襑077Mކ= Qj5{{/f7WQ+YE۞; 8(r4hɝŗTy G?rRh-Jw@SNq飶ěU23>$su=*_|q '—V7v~3wˀ/x<<(xU42aR'ZK(I| yI4HNґ0хT}5}}FY1)f;$ rd-!\0Gm:-c$Vx%QG4Ak*Qޭ6Sʜj6l kO.%D(AF1:wczZ(KGlEjveJp.n|=Ogk ӱ #)TD<fEwZI%PLѪ 7=;C@(@܏af}{oX۠?1nQFh91h#04`AEL"\77`Mm)b7#T* s!ά>r i(y-GJ93!apg<iuЭ/z;i( xG<́mm>K.*kܧH:2.|wٿ+Sʸ') >Y5?it]7UN34iwمxO![S3c$+^ƾQ>dkd}Ag+0c 0E RLݱå/X{JDa քɥCR%t mΗu"M4+45>>^O&2)(EWnN)V|~)ͽS0d{KXji5ȉtJ>;U;kZ>'#Y~-.n^\,/{Ψ-Z\?G8}(aK-lkNl2^Yg~+Ci$EGb|HMÐaBq:a2-@0A1Id|Sz~;rh^G%h"FmZ+ըt*Is@Ng.}zi88GGQ8[Ѥ߀p绗zޝcqvK8,KiMS~m? TQ9#bh)Ŀ] C"3tqOV%Qy3_,ltP5Ew=xH0^54U %q>.WV4T%o NOUݾj&v@'l$IOaS{m8N>I $I{$(lz0.yA>}M#|Yu<4VY0D#0znq8g\0s398kuz9i>(tĞt }]vJP|XУ2LkNM!~q8 FVK˞ ex@ҌdUt ,4 a벒"[Kf .N@L+.vMS5̧?L7 GGeUzp嵙c#6\Ng7&' ODjP^MWIUZץ "\^h hZyͳTE0d/?xp@p+orFm`.W{̣$GT_yDxVv5M ᾽`G7%[ĐLYhZ& R 5P0F:`4"yA)ıpZ+n'g鴨Єˏ"/p*blєb'HD{e+Yf: =FLrb^~߀.u;L*,h!5(mm*~b *Zanop3+zuαy޶9/Xm7mkdc9 .9[MsͱYY)Ao\{z`Z=EؼF^qa]+5Ѷ& ٟj5{㙜FY_0n7mPs ZDZtac9S1;|a<*cJR>IL:rEvwiy`-t[kA!oO ߗi}ݐnZ\QG ^~\4{-H5[JzO߻1>׾.ns}yi6 T޴GZve7Dqq8A)\Cn/vRxY7/S[o#w//sv ՚jg$}`\xÂ0dR1Θmq0a3qP5,}Cc.] QYQї .)vpdg^"ؖ4<] $:|uGeH)!{@P-X=Nrz FTp;<`t-VE:G x5 h!*dG1<$)YL?mafeلL 6B"5gaҤ-e$kO̅>[d mt20 DLF[pǾ:TMtY[jȥZNmՌsfUU̵@RXu*ƪH-tKBFc26V! G6bVȘD9{kΆ)ִ="͑+qQ_к'.g> x ʆM`4JHd"gSؤl(dͪ-+*#rHT5Gc$qYޱj \<xrkT?]g=L 9r'P x hhM?`̗7&`ZX{ b>qe׾$)m{cf^C6]3h 'e.#-D;2d{(ܡI\]r}ۣ 2S  ֍v=UMF (_x@=0`-S!Mŏ*eq8-gvM$%k<662"k(y 16VeқD!kjc2mєl P7kr bPө,,HWL +S )*$ЄҞFMVcT Sr eHK1wLY7wbF Ÿ~J !(Y#3u(*njs*R: xDw.U[ct2XYua0wAd I(o L tSȦ8*:dY){h*3jƀ(P.0є&A*Xud;1X)ո $(WFt0o(@H*ۼ&!FBI3X.+ȁT햡 /\=$Vc+:c"&LLg;ZѮF1"/YfUќ$0_c%*)_lےWZL'AP<w[`Z~#ЙUmV%^E Hrc00n8&8fgtڂ|BG1 Q$bDO W$2+BiO,Q:{ ȋ ":t@2GM8ڂm vES]_~\GmiVH6SRe%)$ CddtݯF]yH!.p /룫*Fe]L),C޵Kp VhGc ;f2jxrI'#T6uW̨nD#+ڔhhM@Zi298kkNn#:]D˄`< Ü@#ҁ8w&3tQ`5 8N*r+/@\0 槃9ml!&ZKo!b!] 122+&) 09qsMSlAHҨ56X~@՞HdXT3- P G^5ͫrMuB2ס=ڻXc(%.֕~Y_~?}0D+@Il{E_*7AT kDi)GnSB#JIEEk'_ 8=TpVPE,j5$`o^Ћp.ߩԞio@>I'Z)_29Ac/ wK҅.t ']8INp҅.t ']8INp҅.t ']8INp҅.t ']8INp҅.t ']8INp҅.t ']8ߪ. tC҅#`{@T`tW hGtoQÄOW᫶_{(Ȣ ԖAPX{Jy,+UJx/Հy*1D|Dtob1V$I f.=^B Qb'gM >6J7˿Oxrf߆t~Nv 36|۟qU[| Zd9&~jwzى?xN.?;ȭE7\ݶw70on<}a=,~z |IWL9G8rXf6ߝ'_=μv|h]:}y=r{wߚ˓8qgZȮԼC|l>-n={uvm~?܋ܸ\jqԭwr5s<ܽפr7u~F6"ϴ`xlj0%'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'r 'rxN[rh~0N-\{0NV1ڝ8Z\t0 wüT)V?Rh?1SG6< A4^D{-vj#r8p3hUwDΣH|<<^GRfm뿿B}<1kJԾgҾ-{1Xc<ћF0a`/l 7iHN /d $߽j@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d@ d3 vgq0mP:(kAyNNa-His k43agk!NsX-k.pߞ輂^e2"3@:rVAOxĺQw[/쀔+k]ޮDZ /1Q! w8_:}8_z^ޢt̫N.F]|(hRLw`ꠒ7pN*8*MbQybrHc{;M8Cpz29k/O'gᚫ?+_O G.d~bgu_s _r4:Nv}9.gG8A/عm Xwԭ_ܛ_~?;r~טe nϦH6z<]3!ēqaw$+m#I.6ż^`zhb^0򔸢H6R[odEHJL^rlyU_DFF.!-&DV^.XW =A] J4#)*HX{w.+atLdBQ@έ!9x)`"3F+B{}Kw}l䬿ݬ}0WѳX_?2,hf~=/)ڂҬMIeuٶMS'ʜC͟lqL'1?w>Z }WTK3t Bϓ]I6}FLtΝU>}w xVh4xTcD42UI.pg , + @@w8q݀ Ѽn-QwϴQ2!a錓7l_e_C֡:u NNt3r2$98E0R$Fdk$6&&nI F`2~ {Md_v@_%]Ξm);՜w$+)/qr_ ^Ty];rbFymaD7mXIjiUVLc4Kf 3(Y.0|άă4'`{MDndх^2`|o@e%!N `2%" SMƣ~00\Hgq2Jp䵦VD })-AQEYªVmU+tިGpT*m!Չ '*`iM+ TD)Ԃ6;l&+U(Ts)g"D$Q= ]P,&I]tA`N2JMt=e` ן[iv-.l>ݬ;Z/}S*^RI?ݜziͻNW-p%^Zu6,|;?js̘ nr2nf:aV$vճZD97gV3 ݃ru H4iQsk{kN:[Uѥ|QE<̇?scm`$ZM}t}^ Ί!]Uvb+~tL & ӫV1xWU4{6WK* l5jXvnx}/_2$OmhC  ;G}нHݢ~m?[NgCwr7z5/|֞ bi_Z,6߃%_`T^sٚf;D~]uMR?~r320{1F>ŝ~f<풓3~ԗ:?-6?rFҬq$.֏t4 mFbY&O!|̻pot4f7i^GxE6ڴV$1Q|&<@>}ad%W3kϗrS 4?zn租˗~P~闟@ v`}+4oO`Ɗ'͜_zkO^syY3>SzYgFNp7=1U?q[bC߈C3wZ㒫.Kn/ ͟ԃ s>r] @xo Ϯ^Seb}Or.HR+-&p! {81 x:Eu8Ih^h(E xg7A) osslKʢs\Ն[&R(a◞,haBM&&H蜄\iKIQ"G4w4a f:V|qLIq+ŵICJ6GԚHR2p8&!2AIoleN`|jk.EkecRyI`ۓ78EqƈW1,H$a !PJ "ZSw1)=r KP%I`UrJڃ`uD%p B_ui aV:Y}r:MU||] XS &P!Geǜc:OJsu:s9m(+AX  ,G*AJ%Z;#Io!{^o\^ΰz"wTZqY$4&9C,,ly n @A3YDTk9( ~V`{)ڢU\ :Zpx.`B{؀I3DZ*%EZ^Jf'r颮&`Y[όr%q"SS#FkuN_ TplSR235.iQ9!Kܶw\GC4ߒ;^gw+(Ur+aPu~>^Yayd$H3x8k΁Ҷ3O*.m<,Ҷ bҖbx='Tcìt84Gjj+բ2($4z&I֌6#PNulo&p}5y{ >Zy"i8> jtךwB`[ŐU}֓Yl~59WG fvz:_: ĕ50 jY'73pucΠny>X>Մѷ.o='~55zN64n{u9g̍Oڦ\il0K囖oΒoIW(MbT+1*%AErg4YN%&+AVmC8얁[Ԥ^o/<\s6E 3x@9m /we3_52f˘c˺ꞘZ82SEDMSI諾JJMU*4ly_1Bڶ<=k xVʝ[qnΪmm]mqGRIc5wa)ڠ{s:ߙX>0:Yjwh6V:'%f <%?!̯5lЬv^ԥ 1RMDWuY=ㆁx6`ūi,TZRy%Iފf;7סwݬxv)U:P. EUᘷ㤀o;Ο녑GK) bK 樓>N1+ΜT4`-0')n1xcPyt /^V6~KjK!ڿ nCOj.Bg/y3L:ssߢC~!OvZg&?%է[у~w~VAxWtkme/}wѱyY' ~xȒ*MErtek|< 9FKs%^YydQY B8YV熳89;ygI~ r3\HR%!٭m#!,ϓġm+PAY6̤wͥٔZOzy󉚷V ߙFCU㪟6et{/A-U{lx:^ u4Kf o;~mʳ sl~GWvh|hWS99}Lͬzi1 ėP^)헗f51o_A7645va2i537KgCgyb?)%CPdgozE N>W%VۉM+!JLߜ/̻5άoġMvZ⹫ճ;8s^7h7a?=7V2>Q)np05sHז[]G‰uXTqtqt!>}؄J7Fh]X_[~tJO$*8ʝ7\ 3f4j|lHbaJoԶ~nSn{sq^> rg)oC\jqh$tyC|e|"c7s`eвÞg{$Y6.kZ0m>W2C:a;6w2^k8c`Wr7#۪m/^S VK<9x MYd5^5;x4μ8GˋaqoA춥_3Sl& Y3XSuTVm>2[|߾J}nhf(a733љfcQ%kXyFoak7j9ތ#q*չyrڇOm|~>ҨI=qSJm"V@kMN:(g|7@~ ,Cv Wp5S%C!`Z>4s ZM+%}h<\A`BbEԵ^ՓP?jf"m^JSwFd+3CO(cWwzŖnAUM;O3ѿehX ,z̝vX-"ro03 ^MݩL#I韞iM3}*М\}cd6Wۦ͓b>5JCsktJ퀵[:0IA8 hTfb +Lһ4#χ^[M>as0i>Yz{Qw@Vv;t$CĘYC +bԆm$T0NiNl,3&faF:MD:Rb&y,ViT`Rʃp-A[/5^#:Nl8f.#Y Ƚ4A'ӱkC7_em"-!xn*l2-灭NA. !:F 9,(pL:ĝX8m5xsBCQ¥!L|eÅO')@C.J*#R"%1 /ff*!ZGbp7ǼzWv{̺Q2ƍp"Qa4ZNl L% XPl[1+-=,Y+T&v* AWgV9QI4qżY#EtH9fsg<i0u" =u4I}DQ|nG<,`Y QY>FZq/?ô,M }2= xo^kHb&2`.KW?:639TxVk bjW`%e.B?o@OG2ž&+8)0F";BGRP5I8nz6 I6`MHtHv? 00ჅiA_ЗΪB$ߑUe{v7VHɮ.rK.Lūo'SpiF5SufwW)POM˛tz{'iAWum+NϽN~EHpңO'*=aq2yC_MszcX߁8:|}_Ã~x4w@img[|鹅0ΤSח7q))l~/ Pp~vnrbUJ5qq F E.N,<^X[W#}߿}h5Γc@%"L #KT68ϲ<7EڋD&iYN \I:p!7 &KBh[ǡ[ۨ#hWCYġmc 3i|]s)EVUY|li| `}:xjsO *GF-V3CM%z):SsOI\ZwCN"WsMV `y"@ '`o*s;[Yp/H,.db)#Bc#bjҚ a* PcQiE ]`jNN=2()&8 Cl8-;;T]nr)Nϙ78X"gsm֞ʈ 6AIσ̗xVwIʗMx83DB΃* uybF`9WU k$VFn ΂dT0*K[XEd*`"\B6xh"B4V=s櫒v8y9#Sì"XH&(q&9Q0lh,+X*%/MmV~OH=3qGP:SzG>=" Juh[xzӠ0ZLÂև30av0+d{XQO>$Ra15h  tncةbPǎc'%qlOUIٲL?[f0\~/\e\HWPxRrtZy㛶ߺѹ*M&oP+ۖF@A^w ȵ9 K 6eyepYDa5q7;#R^Va4B!bȑ)IW 6ҫhRNx-,SK"% e<Fe:L CY'TeVi_u}V]޼ʢ Rq-zUk࿉*6_Z( n_[7֫tf E\=mܲw,7v>Χ ߞO̹r^ӹuXy#) ыJxR&^>Aժ$KHYڣ1F%q1xS Cx b\RJ*JEIYcg?ig=\Z<{hwr>)&帢w.ǹd-d3CsBKNr3-~c昌G5{{ Ճ`Q ?F׷l"e~eHH;=hNDʶf2gjߜOko麖 VTRv o#i?]!otujWw :dS$RI&2^mHihfrvdo 6f2qsssFҭWI$]1@}F&SJ]Rr/eޯAy[l-ltQ.j'P_Ů(ˇo`<3wa.d; <Y,i*g|89ңHܦl:#BlR(>yMVLh2ib 22u|W*ůa Zr *T0XCGhpKYl.zf0BA i{( (%`>l{?ڞhFW-m{h"ث^5#ثVvtX`8)khm bvt3"+g٘ *\ύ cķ'k\p|(GkQuþGQ`ߒGA츸7mua,EϪ4Ak-R1So,Uep}}Fd(EA҈TGJlQlmխH&UyV1CJ2RNKYM1)yݱ Rqx57B~r&6r>#npvu3i{ΫNgmVg /l; +ذyQw>n\at}wWߏy5Qɮv4ύeB]4tuhۼf'nVϧgĞ{&9-rQVlx.-ogZlvlhV)ٿ6N/C:iZ^dGlF 8RcaPdzػ(g.hY*8 B$_@B\Q:-M8~܅QAp kz#a!.PeVUatY$\$|Ǯ8sQ[p7@fte*a]OT}wT?5C^|f@2:{=6IQ^tq>TxAժ$KHYڣֈS N^$.oxd(SpS,2KQ Zt2X(бw&z ~zz\uWZzSL<_qEۏ#̰)"/CtUW-FőFC1bɕO.rP-2.Q8ʂ+ fه5Chwv>>v03ѭ%k*$%Ơq!A+,zp -emFcBSY@;e;HD 4 n* ~e_~Ѵ"4Yɑ1 " їLZzJ-($HNJ,1o3)dAA@:~ty_̐ hIx(B&Dm\hu_SZ? ꢊ!}^V7!~)j3dS/tR qk~y$?ԟU[];:-b'^ q TǔPRHFY !\-o'e{w?eؒ! Qb(›om#_!2U׷tg޵H=>-wO}lzu}jol|\Mj,9IXa >Ծzi6 +Fge0C+?̃e%Qo8t3:;H!m+޺tն k[F"ZYFhcj;R|47zrf?m9XմU/ٶjۻR]:A'G.,}6fzCqFި_'m[&U6^__: ~减2H?LS\6n#+bV({C率/$`5f>tES(_XfSnc[lȿhUjiXo47b )34f]Q[҈;bB ɂ*ߣ'Y/GT '$u*hx )*ruTx; xvFK6̹KUt0/7|GDq)0(]6*,DX^ s" j%x:N$MoDj4Oh(. ˎF O;9&!i1_:LJ)~Mxy& J,è|2F*fE4#hR_UD\ݼ v ^_0ӁU1e-0,u:h=^kYlNƸ>IƝ'e$́թ`])`|C$ $&/-P *o+?N zp9RR/E2q1b7$*=F3eTy!i*Q&g 'ͼ򨿇*Xcq<3cFAdILJ"!hR@VZ45g#UDD2da$ #FP_ʝ骋c:N )pDB567BKfX`92W )RZv4vD46N=tέ.?Ѷ_uסURpsBu`O+>f-^ \5{=u͔%\g:(hֿbI@Nn(CyYc+ሯ"~iK]-}zL-ywr.x߬|[bIiMD9rx JY0)(ʨ0\L4/zsɥse2XZYP1L`7@=ƘgZi3+|sY *:J?;xū `Eה~*]u!/^Dž`gRlRTY} x2$/TI+y,U2F4Ny˶ʣ+۪l:甍u2LIK1)ƇQ9<$TwZ`-8tuC%[Q+[;eRmK[r*]uzGuӒѩR1SuJXq d @(;]t΀W d["}.uX;?k%)(-`3I FJp+|K: Ư>fHNw1=ӅA5]\XiPكK@P~@ei<7s=n80wjr;>ݿre6}}y6qǭ_w!m:{#z* ǿ_/覣E*(tkFF0zYj֜I %Δt )FβVnPwJ xʟ&t֘ Ð!ּS ?UH0|ewevsŲ(LXQ0Q@ Vz;(J7$ZZm@'YHږqHӳAd~Kӎ[8ct; X:y.%Mn'tg>lLu;:yf Ga&>]k ћgaɆ9y7#l3b-4&1q;͈M̤ݮYh.g S<wʪJK(t׮l24/ fPwUiQ|[+NUT$ORQ(-#KJ@!Qs' KY4yρ.0!d揩KA1رKr[.d!TW6hCKS^wA 2DIk0hgj*#9i(/sF:Z˕V^z[e&yLO)$5ғ[k(r}!E_]RBJ\E"B8K6s}-h%TIn<'[m=r޳HX=_.S. AO= >_ao PD~S6}߶ooa?֏frtWuva/7x%&>=/3.PzIClPeCQٶ5[KN;ǭY*:NyttuWC܁0Ӿ >7}Oǣ|ZBVM7ghZ}Z) ГI]2sR9.u-_fϚ*p'Ϥ .r}25WW,tKϯ4 a]n i{cзmH<^"oW?^A'Ǜ{_WkK?ERR23C7Ԧ=kyoLb:S#u;˝^iTDݵPZ&9jJ~yL,&K+d֒ T̊*I)|*NbS׷.'uaJfjL6[ZT'sn zdʲ[Lds`m%6RV*zo2M$e剠`!hj"S9) (gTF2`=)`z]>@~Q*Jsblwyr^X 3v兦' Moi<]J*Nn'a:o~3r?~HYsó4s= 9W:(ᕈYs#f&lV2 ڔV.[Nפ,vICP&SI1Db@K"YFJڹƾ"qI NmVHH>`DRIQ͓ q&4w%h!QZGi='4=QlHnkѪ: wQg4:p.01%ipM*×0>&\]^YL$Sӻ&/gnAmi.&fl]%xTIVeY2 d$EsbL,E2kdKr_[~>#b&-eK\]~B}\ECiF-t c]]RiA!mڥzpd~d5ۘGקpcQo>klu G=͛?F0cjLq0[bhì/罫l U8Kxkv l'&߾o#a_?<n+8(;ح0p nE:ߩw%w{aKo{q "pCq50{ĴEyo{xH%qfMJ[ڠD[bդIkZ쁘"q( W v9 O~ivL]`i)ufj7jt)d"I^%dbڸ`EuNd 'Hh!2Z ASI%T$`d%, KRxJ5pa1pglb<+|E8ތTYk)K?-ah?T S6DVf^*Բ>Kx(!lꪵ||" fBh`R^r svh->yQ) b(X"O,QiLGuY #) VɎkƙ"'ji²`%\(L6)iO8 Y1p l^nw b3DMLI0=<){%J8$ ]t&1;(I1-_P`Qed3Ŵ0cFP#[Ƌb qd"WS Adʫ *rinS. >향Q‰6@@[R`V♊YEt"Z0w`XbD&"nRw !GcM:OZ0-xk"&D{ƍlEK0Ю}1\Of2dFֶвdHJdQ%j TթYb1v~m k!]J"@mK1xyUVw %ʪj'-?߷FhgU ΘkӘ #}Boϊ"Xm̊}WNT{ծ\g7& Z95ZqyY~m\OŁ. OE`ަFgy[4꛲qW~?\o8Z:@Gu> ] fژ)QRGo"KY#nw1̱'c@e_&M&ڊo> Ar&q]?A+mKܗ/1DLvI&`V1A=1.w&bVM9ޔ\,}nQ6e^e31hKYV.rf3-Is)u1\)y2ݮ[@@2Fϴ,sJÃ&f< 1X6^̹'{ N+*)qs^A_ $mϟVO1^2ȹ>M魒zmKnS6*Z_izFhufXwal&ɡ.< _9n S m!OC@J}!P ђ@QXQJ ږ.qH;&Av*sd\xV }b#K1FGCg/1TϼZXcʞu dvdƊMcΑG=r7tYBMI^yєᣅ!~^o6{|d[B 3*c!d#"澋BꢐdcEW{ @۵$*,uH9(8Unr&,Rg2D`zLA p8%b"XZSU.Jan9 5pvG!OY!h9&5 =s`΅]pZ*`5%*:I$8xfSN a(9=2Be+,WkT_ّ1SEh+.p[ >^ 7 #^ZHh77X=~f"-~Os %~Hw-ӣNa_ITnJ_T^??Boo>6RJw6~ E5$0?D!ϝV sAzwgpyX`- 'y5j_*"GԴq_tv([Ajm h4hZrE:W?\"Λ([=jV7p៬iu: Yvg|̤JbvХn>h<Ǭ 煸} yll"}4G50?W_:|x{Anu7 Jlm>}2݃7/ FRjKx۫/Gج鎁 z\q2zKjΗ?iaвSgtwZu %wHF{rlsJFV/%.ń^z\D..{X?z-XDKkD`KUkٞ`R!&p5zE>Q3RtCRvR)|;Wcj\jH{Epn()Ӂ@-ÊcaoHԧ"G"`"RSFDLM$a$+CʘtˮijX j|} F`h\''\ToڡUyT4?ɇӥLkI4CgiNLG2x@b[ YM7Ǵ.[{wT뒊bױ딡Gp_~Ի|3ϡ~ex cf 1kEpd*#)M9eVz$1HfQ A2$Q( | \KPw,X8zJ~#1+YMY>K9kH:rG,l)=r|7`2kG|UgĚKb2&_\inzM =+xaԕJه/ & gPDbtt * tV> I6`M@ZX.*lP0o}ĽOE/C,-_#ءϹcoSA~ҳ=w(Qy?&VT+^OUS<-ⶪ=8K7_8+VDh,Of[l fTDm3MCiև(OI|9Y`VU;+Ag/nֺb*^:M;aN#ibĥ76۟>- ?qw?;z~571Qo~#h>SglX3ηw2O]pvt?N]HYeˌp랟_#rml˓ ODSSLP.Ѡɼ%ybKMK1Jǯ ŅjJ*l/TվjRy_lWa$IOaS{m8N>I $I{$(l3Nms,'oqsċ6R'* 2aR cG f|ty5'+\wʥ4omg;‰Gjf^?-@,鼧;"6U]yⶨftY'yQSk9C? JdnUSdt_v=~+32iQ}yiVCP0xZs\~Ә>]e5s$o@6*FyOVv9fˠYP8;,p;` zeaJhƲw%]X a#>QFb0hߛn#w~M?cTFJ/ _i&⡝s?7kzJ-si]C0raq1yy  `r* $p9})B^z]Qq||mVx]}K:D$rW)*L1X$Δ!.g=Uˡ"!Dq<ȑ 2ZGH*fs倱ZHsĊ(Gؠ_+ p*KwXEL_!ݬ_(nm}J쵣{/2ɧq^+TxYG7\$`"{&XاLD$Nh6A7IwN ͣt?|uP6VmY1COr->֬gI߇iGCQ#eQh ]]:81Q -ED˫J6NzH$I/; tYyD^ vc0`@w<-!?>HƖ}SJA%`t k!k4(^(-9m<%^Xn1m}O m_nY8aI<nd~~'E.I6H0Q"-*@xn]I<ķ8.18Q d%0O+"3u9\/*3D~hleJYs}}/~l|{W-~_ŗw^+ASM%$TPUJBUI* 4Lhnd8P\T8PZtᔥNo $fLpeér8UT9*Sp"B *+S 6+SpNér8U.ES9*S՛ZQQ9*StNé*fUT9RTT9*Srér8UT9*SpN5Nér8UT9*SpN5~jlo;hf8Gx/lF y9y|"hai6.5ipIL09u߅gjrE&Ms9FCwK:l%(at\`RxFBDN ăCx /FCՃr\ܳ(x|_0 wqt9^ >ey>!6$z?=4.rkkn}woO1V璒vn9>as:> o_8f\l*kHe'75;Pgxʜ PSBi)13\~}3hcLoz{R^pG*Fi&2q9C"e`T<j[#Tb%+W:5񰗤(¶K"|_Gccƍ")9cbz>썀p#xfwAo!Z橏Pok (N)}ᇽY|nrٯ&cy[llW~% 9ӯד!S҇E˷ouN/ɛkS:ۙڇҳH1z?!z!!"$SDN4#@U^K+)yzS2JR9;JQ.D*$X1PTMhJ[b⬷b;c[[-<)[(_m~69Խ>gO;:74rrwy1-ݥj!?1βRy NBP`0ph F&J-%z&.jbu5-azQ@yThB$^ȴӡ IFem3''}$UuC3" 85I*t7 Xʊ*(nsJK㛪iUӪSNZkϻD'稫gm9+50( `_??is$qEMgB  %\OHIyX\~𓫫ɸ2ºF^~A+r3oyjjgqf[T16Fw]ݕ>shk,&4Fy$I4fizj 8g,OBEq_0R;{0-4XŚ~8m.-q=擰Py2KeTV+=)rEIc#GRi.o$L:?NH]Ou42QǸָDh-+`$S*v^\]{۫KBf^ s<("Vw}V1A0a/K}2Y\N0KkZ_ 33e}4&_׸%o=.Tg[9FMq__EW׭%O4?`9ݵ(Ӯ{'2jOPGgg,l$79|?C2e@ON2AD€9)<! ?QnDo!Ѿ-eĮ=­hDa9N=RPF%e>*>-"M%^jhAZ!A&'u(e: T9Tpn1q6(&Ccb#=YT5يcs)nCrP|,G3/}_2{h(b3?Qv6Ig^DJSꅣNpBgSH(k4L\VXKhNSb4H zv E) A)TTꠘ $R4.$wZQ=$Yg9R s@Rp=KIǨ.rb /gPZٜxFm^A6r5!^ZxD\q7T][,'M"cœW( ?P4A 22)?U{ʹf"1f6KC ZiW 5"˄uZB-sYqRI =nmCB#`hȔB- L%M<"9]"[5´<\7qQkyp'Bs)<@Su6JCW@$W܉v;Mn7hx`mcS 4Vh}B48 .IR 'AI0U=ʩ";*Ƹb;d[(̷[<ep'i!ËǏ}]s8Q5Um 3? 50IJ0`doúʱ JFT #++"@jպ Ao`'A`ZK߰$gQvJTV9D,% x1(3![ ѐ_#ucL  4q0;c$:'-uN \J o#: dۖo2'J[x[z曞vUy)U]k[o-RaXEGZ =)gϾ|~bѾ6H bneÙm$%LoшP68*J? L[g6KZi= )*rJiFS!Lneb&N3[?tn挘F ܕD P=|1͌}&^ Z KUJb%X']8K"C*=Uy(o}'4 ,Ys[kQL|[́ @ɔgq;,-g)%o_sc…=GaE?7͗`Xi$(Hkd8<P,f/=چ_ܹ؅ uoޞ ỡԠY;vo/C\֥8ʙ& c9Cb~4zi\ gg 14܍rxl{?6.yy9b-eum 6\rSg[a(vH9fz}Y";S]Xd p4^kfE4 Atu{)CۺfvQV aU ֆUY\i ReAN= Y{WS9kyg0BvzzоsZGXlr׆Ul)HIR2bWoIsl[KZGBl}(KKGa C6QI2]P*ؤRS m̕L451|sD Dy &g He*:VLc-;VQ6t◷~vnvԽ|r}j\),j먛kq=|ʐt G78^|xzŮ'Gϭ0އ|,ޙ+4.dErJ8CX"jSs(ZK,z ©(X('52Xj#G jk%NR^rJ<)6XLlqioWE󘢃ø"!K`Gs˺ៅQJ tlԠ碍p:X1š7( ȷWm9wkоBɓZ7їS,Ĺ2%c3VJi6\Hj85B3y.4Ja)c$pPp6r. ;D4C&.%ΆxϏz&g!-:ߗ9c&%YHr2 peDzemIHluS"ެ wb'uФŌUkZ]/$!iV%uҶ`lNiXY8 LݕcP4Ec9F=^!]E!"uL0NT%j `@YE`Dx(/cs]2O\vABc ˤqU) (JrcBzH+͓E ‡WgZݚbV.9 XNEm <\nn"޹YuQŐ>VFyH_?Q}dPǼǟxV jW5GNj? >q7KV5t;2" `/) <dž@uL G@0Y}9͙7{, $!(DmnW*6/ot:~R1 >k{IȔws;H_e"Yի رbWGiQ<,M+ziS4=T$+kZ,W#l]uZbayͦb 0Dp\FUro7DdżM|B!ʹHq&7tmmrqMdyDiLS}Ō2l`0YgoƉix/̎N޾wO~9y Hwޜ5K)\m h|46V_REk Ӛ[矇T~YjQ%?e[iXEq%qW~FEp6HGlmfC"}yS;媗h˼咷j!喅8|d~@Vsm>$]nS^@?l$,}0ADš*%$=6ImXq%qXqoJ]dS߅" 571ƬCq&!2j!0#x-5=N':3Nځ?hŷ.t^/u%~;By -sOQ8";a";V}7VP ծEvVc";z 4Onyn*Dme jKofu&>uF2ogVn }+υ{ty d)kH2r6#hlYLD ]͚8RVCRXzAJ},Jl D: r%fѱYp}i I$r%zet־%) 'KPX6D;8xN(LW}aXJ^{BKQ Z-]bQfkWluOcDcSZ{:|ΰfM}czf/4b}j>wQ\٠ 4>8 DrU9MV˚=Ԇ A'-BQV}.6A(ag";[ɵ6lgf"0k+d] ݚutHE: ~*ʽ^Dh3cK%0OivwmMY};H]sp쵣hG ŶT-ؖ 1- c[*-?`l N<1wEx4GQ:8aؖdh i/:\oזsT`[tAŐd I2⬋*Al\Nɒ# hȋNm9XI9"2#7ZzZ%Y1Jd(mճ,ؖ m9az 7SS8eVq44n5s!jo= 6+Ta*Xkeޕd(c&` YG3Rx;!ޥ_i}7;zf\+Z mٓ;/V1E&#ZY6rqYX$Ծ9i.Nh4o)R˘PFK7tc}@8.t%hNhM.@)mqB 31M:QQ١`YUVEٮ$474 >}by[.iP Dh<*2?Lf; W͛>Hm~n~O>OWoݷww{0_!- X|P2wsQUY7Ww2iN TC[nۦnU=z -.ڶ.;jP-¨mȻ@Ys~~vևM*Շҡdثv}ŻxUoR2PkfVkwˮ!=mRL}{?]"tE(Gs 1z@$Em! ^Gc4EߺOy^pMɏj;p4{LA4Q!tF$R)Re"#.b&&H6Tsd- 3;ke 2A"L7V*CGhpPѮDt"^*WCxs؆I OX$Mw\$M?4h<~inYEo[ɝqrކ G.jCOf+r=0C0%͹u%DmL6 ADPM deiJIprFlɹ(#8pdC%m.>(R0XL~*h Tל3pv(Dq^f<(^m~Q[7[݉/Lcy0M>wαv.AOV TYdj-k`9+ RL ۄuN9r oCBB(lS FJ{"h$*"Y1 9vqy(\38zk]D34 AYfD2Ku:MD,ۻdw|gvI4U0" |qf٢,b!2`lhU hLm9)Ҟ^4=Qo~s:kYNgՔՔm1+}~p *P/L@K)Pu_&B/ dW#o Dt>f s*]-ޛM|e8oZ΍۷'IH6:?zᏗ_݅pU=Y?m(`ONEыpԪݢ50q&&Ңa+MhcL<uY4!:s.KPxm[AQu08f2LJo9>ߛĉɃLZhH6$e-z>}/d;=?2|XI 8WJ& ^LJQo'v3v|LI)aa+ (`PZQ% $R9qt8f48?pɓm߭"T&tڈJ(bkERG@F9J XuuRCĞں0lSjj{~So394ۥ|2)$RFe)sIB 4e %dA7DC1W"[# 'cTe0q^r?bVk )fPln_f?D T0$x%)-} ф<Bh3 ;1fZD -o|b(]pQ)M&޲@z(ً]k]c5-j_ ;ה| TRILCl&TX;YHQT$$<&.f05ʃa>^(ZIQ',L btd70XY!/{ضy)C63; pžlj %E8ϟbbVKܲ[1H2RSdb]U]a}/?*GGaC3z6 Y0>۟&9*OL%1괰̥h\H%plJA Be0۟J4Hu\m@8CLg7s&5w&΋i} @%&$:Yx.#)xbxNH6'";O&YM>.*g{u:>u8wT;]k49.gzѥ|1WY:Da^]ɏgV[g LRT8ў"e{d}MsQ@vLF 儝مStl:,LxWK( 59j=!٘kPq#"7Eyj"&1ƶ>Dkii_} ~}'k4_ \`jP-VSO('.sɇ[5*ύz߻3E޼yv1Uog/ޮ+fmqeiп-7k$ G?gu@[~O@HH7bHWMða$!8fyLpTOy 7˅ގٿϚ'Q>rM6͕rIjiK >MhiR/FSP[:'=??_8۟%͇>ooPW`}DG ذ+j3Wqq0ܟfy9V 䫈tI%ֽ ҏjhYqY#߈m -zY/>BSnnŃ |3]#A}U7`ïqGwBj$-v\#If9nF`^ʕd6Pp mv= 6S6,OݬNEVivhkV(MZz?x}!=F0A_.uZ=9yҽIgDrB 3tDsJSH 3$~ 9ۄ[޺E ߷41IFu"J[H9" ;f:̔ ` Q 3;iU(wB"ZSiRqe\[Gz A'#Kol@ P(I!F Rіm٣84 :ä k8^ XnDl29KJQrG!DB}NˌeL  $*9OI Di$x:C\RoP{ֵ GbHNgcN3 ".Fyh ATYgD n`*nV*.u T*a$sPiˀZ -'& D$΁E{^{'~WAZ]Zlvۋg0DUl|)o/}?=_=a7. kȝe8kdWcS [׹KL\ڴV:$xOUN~XZsRr.hCL$]T3&H|2 Yw?cÚ`wM2կAx{9} e%B5|׻k\M~n=8}612P۷"yj_ &*S_k5[ X-hZ@S‰d 9K 9HʮD2L ĉ'Bg0 s*j6RhrjNW%Y Q*5"Vǰ[H9 sg+EX {fߔ/3DiXTx:z/z_!/ϟZtDuʙKxW^*))ɼ#qD\Ka=95!D3M-Ɵc@F<Ż}4an4WuLK\J-Y{ y*yeI LTv$ݰtv%!l3Ҡc\mk5L, DB]+TݻzNz7>4"WT"YruND!-/y:12ٺix iM`"QX1fS Ϧb'Bt< ^Pj)qϘ$> @da$ $.(l1JMuݨxg%_&i<#G :g(@, <j[4Ue25rD!̈́<i6D9][ /T]hkf;,cj?h|k{|WކR7Yuݝl/쎮RxGYgc|,/[IXs ?>e[uc [Wغn][os8[W[VزZݶzm&Gk-h9Gô2f9SWҾn]mMnTwl=z2]o x鐝[1>Q优_^YZED~q©oom!'?" 8 Wtv/{lIB[!rсYF&r4ޓ4~ՂJT]lvgR &N &dHfE1RD[ivo=mÓs<}w@Qr' "Nlf7]5އ^.mlϓ4Q%HN gHNU Gܩ9JX-%\ JB@9 Hvh(Xj#G jk%NR^rLk]4[v.شu[cs&킄TY; $PK hf Jq&82 %f,Yֳ6& s)arEnsosrt; A3E9{1I ?7g%&hO _v/kp,Ye]8Ο S<1Sl-YywSÖWz!v(EwԶM2D$m82O/=,2SK=x9ws)ſ-Wtzd: [+)n9cg[Sd L"2aqLΰM؏y=ftM2x؛=K.Ӯ&3mh'E:TZRvO*@91-jZnǼ-9NZ<~`c؟=4FI!:b 0#J9K}q+@{Ny}}}p })/Pr!Uj_0f}a/P/Pr}}}}-@ Ծ@ Ծ@ Ծ@ Ծ@ ~WDmAt6W^whl2*iHQPf,(Q~=DI(( h4ʘ8XCxB3y"Zy ZUbMjG#u o{,XiPAԲ9sKeajTsAxm>اV/]]~*rPZvs[5w}U'T4.)J"^4.Sx)) LɎ yN$GG@lpLWIYq)gٞ>fL7xL77]6;OAH)CS.RPFnwU2X|Q[Eh&8Z)}VHI9vdR88v m3C/7w7/7`2}_LY•[k= ,?f}~%楖;BqszFҔz3Z"nFcpѹH М¹qH` ppsFXx@pQ.,S4$A1H& I;U>$Yg9R 1K Q)C%TcTZJ9g'|`%jF!OK{ /{|NɎ% |Q?{϶FmqHCNvv;/;X`dYJ$).-v˖VSd]U{]KEVʂ3z >Iz>:֌|./RB 9,F!ѐ:9AMAL+՛.ǃ4Ik=YƘE% #X4T"aAJbM.H$ 8p≱-}j|x5ч)T%KљWo4 %[SlVcNILgaخfB|ȡHfk}&5դ$+):P@,V"y-Y_yGkaIӋI CJ3uh׋52ۉxHYgkyL-yTcMxї<]ze~;rs46rcP^sy22_[{\!q_!>˧"TgNz.UZ_Uѱbk@-nv;:WEEpίx_h!Zqۧdmdͦ8eKJc\V4_/ W(-EWF?YT)CdqJ w`h_s2kŀ&/㳗TL >; lɑ -y/iHBq#ʹKXckH ҪMm]-S)Vjw(oKE$=sαtBCڽaǡpm nxֻlvgUB05XYDTԝa F9 &Yti!@ W~Xu!E4XCfPkIgǼHVeɲ=hC)FPZQ/4 Hpw@}]O}ßW"h"d(ҊJZr:bf3?ЩWOR^^G%Re!{(nǻ^=YE5e-en|>#Hs y:g17*u n_G7ћ̷Ll' \wv~3/;|kf|M[=\kfϻNws9)~U }u֨j{:*-KߌyI)nd9p;m[m厺< !|]7]}p}dEAM-ՀVԠ T-AwtAw?<] L= Dfꢍg{VR@EMւ5@.Zb&5([j(@e1K$a=>9/fӯ߃(ߗ@|v[هN W1lg'/gcQ%^(BѓM1(Q9x0z%C1x C A& ,c S$UIpГ۫QH?-z=yt[jp2 Jz-|p5 `CB !BtइTھ1i8juJ @GDV6RV;R"9/ yAг8;x~@jv}4R?V>Epo `x7Kl}Ge'n+|SbaSʗR"BF_B:$L R䣗,栽34.c.JucAWZC]-%lV]8Tm`kYR:1b=cP'yn5S\Φ@wuBNA2>(F#Db>Xj_OҀ/u5BHEZR" X0(VA>S1 U*隸M,ZHτeOD ‡b򫳭l|tJlh, SU˔1-nZbHU"_F˗/5Mp\]Xe|ǿǝ&v蝹.n GK{51wGVsmxu6H 8HVSH"+Վ.R ut!NZt|GKfGyhEWY*,ɐ1Zw߸$U5tv>ɧAyrYg > iNztZ&Ҏh_۟W-kлwǗǷ?jxӦڏNxl fY ӍU}rh1zDTœ.O|-} sFoo<|=l9[Fd:M>qZ uJ`^kյ:,2RL['XOzu5.7ǫbJl^uɮU*~dF†YS_hZFm#9.I q;&\Lv6Y<SF?1;ӇS?O~ ?}~=8̩Jw` cw8XsCs̸=Z?~~B9ZzUֲuaV ?e]^ŕ'4W~Fl>Zu<c@߈}-뭖;,]/`u-w!SBP}tN]6៪1XT 1?H`P,䃑 )gjtK:d郄X7p'1q,AIzikÊ4-qHG86h4#V1 mCK) JJ@ QFb48uzqšG 'X]ֿv>So w\5[J,/YsɉEx&fH Iфtxin|Oin]$Oqsh Qh{YGi}v>ftH$%Hi&v| 9|]L$Ea Q8@fLJ8;tx\WY( e4k-d̊@d 4ưI"Sbrdr=Qoߊ Q /D|#0$CVR1Ո˸2;{ )`2TXB6"bD$P!߇EH.]`G{I(LWCaXJ[HBZFg2 X9)ˌׂL@cDct+}ikm|4_`Uг+Ҿ CϾPkm.N: ɪ aMuI3d}Y-ɖ/X*YҸ[fWUw}-4.WyjqqEtZЁw9Js n F[ys FL1W/ӓch|f}'!0kk d+GPLH%hcD$SQѲ.HRIS4#J*kּ9R++\IRQ鬵 UE5#n&u^J>x7aͼۢ}ʚFM2ް@8UB3=*K$8@@%%ꇮKB/j7_/l?^ާ$ΞGu#j sۯ?1NMhwd'SN!#oʸ¦ _\XwF咍 :.ut$YdǃNNL'uz\ȵ :Y<Hy뉱(%PAA졻GQtg5ΌxA6TVs1@#AkO?Il&U\s6N:݄BtwH7QZ?(hVܭOۄ<gr+}'0s."CXMwxp,mTj!<5+>;mПl9o$%@/?mc͢Û=]e/%2UZ.Z6RO`f{_г:zfN|͜:۫Usz(yQqԍnCmk|Ʒ~هKlPXx!]܄.&>ň`qU|RNDPN,6NiPTx!Aeou6 TEWW|ZUz]#Ѯ{k(oq .[_u}9WھʫO= e]+7m0KlnNn)&s4 SeуUNdUP+BJޜVŃQ:@;/_/afM]!r/_mR^L5n~5ue|jguʓ.O.{=X 9`3'`+Ȇ5D&z]~<1PmgNJ7 )5lBFm Wmz~Knr?N=Y#%Ko`H`K|Czkg18ʵ@V#)៲4Y&i_&e>*UVAklQzhR"g'&kǷ'VoWPp{mzWon>1D} n5^]`Ɛɀ%eT9lŁ m!}1F?+1;GoH@3>hX|kg9ybE9/!9R|c]:a܎"cGZP"Gj``.K(DẀuE T0>kl3J.Rv).^=5}s"3AgsGCd%Y6)M˜d IFj*#%}y Oˣ{/FuH|7S>Nry{}Zzիn>qTe+5uSݎ|zX? V/5ޠrK7|?gP 1{^3Te›2bޑS9{Uâ}6MhNk(MgpX;>k;usn%~`M¡!tϙ(+lB-՟I,U)9S$#rKN{LPԬl?T>Ax <$#CyY6 YgsXq3ZFTHRH%LFUm) wFYJ٫(cvcFm-<1֌Vl6|5>"@,ɺo!!}ǟ3$'լG,We xHxU9^eOS7/@%F1tI1v>u:0h%WCc·eYKzgN]vBAgP.W+euE8J2g(I%Sw(yLlv{C#Ycd!@6k3dTJiB`g6g 27ڵ(yuaWͪױ9e+/}88-z@b5WfY˗/T' m1"fVp j!DT>J(DMfetNa!d{`,6n^H,D ׅ2d$rNDédTH"K IOk6[CҊYZ9F)qBS KaXN7flig] ٢EjR|Ш#@<ԕTgD~ƃD~"+UZoEz ?}dn=!GF5dcbZs-b"Ql BMl+jt4ICto*o"$^@Pó*)%%/ _"Gr188qL'vi7#ml:,(PΌ_$, L%CWP% nEe<]-/d򰛩ȩDf J% %d #Dv< ( Iɐ/<ڙ=;-4dWh{\<kp>O,>Do`N@2Nlr$ (9d <|k#  tAÆ&#@@i&A (zK!ǏfSv8v<c\WiQ9oi |N#)}`"HhL!rcY*%:XYJm3C2NDjgL- )-5w{u<'k*W1ٺ}QGKI'Ckg0ZؐEM͗t?|Ph4 ee Xe.Vb`tp0"}[O${= UJĎ=oGXV~Z]P[v]?{Ʊ_YKoqҼ'MEE a`Y#zڲx 4wg!91h#04`AEb&Qp๴-? .G4"H OiTye qfu@&ג:p]GUpx &r/`Da[^ʣwH $QE-4GQy&P"EbUZ[ì0b;:nVrq#)4kY*U7ͨ46IveLUlBg|xz1?mf7ͨŚ-.(g%m^v#HƣyqIߊ0:dAR!]ԁ!u`B&0 (Vhx$l sp1Zu9:ucT`:KGgRF`xR׋Y1Kj;8gkuS&oTOYv637;~q?QkXq~ ̃K֡٣Qcwpsn Rz}\|}=o3D^.7NJ=֧Qܐ6 @[g~  pQPi`yCS-Y#r >>#R4-ASvP`!:\zypP!JT#ńߦo]&W ?$=FtnFh^S9C@2Ut3K7'`Iцvٿc7+<4VYpDT!X빉J8=sq1ϰcek7u٫x}γU6蔵Ůj(Ia9}RO2d74sB >8(=<(ӓRÃ{3;^D } HV6t XCWOGaK7&29qXbι s3(`d68ЪtVk)]ze"*'I: -hAL 6 ^éo[.G%Wdy ûښXJۭ~+_KwVvޫ^T]XNmVمohV*Ȏ~5Q*NtA#]wVU 1old:rU`\ b6GGX##B*+K!tK"RkTGQ OAXyaM;Os0ByKNiz$' <'"@ Y^VHƦLh-']-jb{0ݽx d͇t$YwLN; פH IQN)93Z*yN&zG`yk/9$qwҽOoeW9]|gL!Tڃq=ݑ&[cJE^iGjdY@SJx->cWLϺ4+m9t^/6dʾQ6vi+ED'ˮ~ D. |{}鯛rn_F5 mY"isןv/ܿ鷺bt_5֖Gvm-'H=UF<)c9ϧalbT3|%gS4aF׏at5Pe?'/FӚU/¬z`G͈2|fvTRښTOdRqD7U[@lRo tsIM2勺 Kަ޾[W:X ukU*hOBA7 nrBLBul}o-ɯ{ ʥгfr/4Oߑ'WicPӄጐ}s>t*{yſnwh+l\7D߸[߽wH{EpnA()Ӂ XjC;aR/1",DDꥦ0ǀG!eLDa߽1tvݧ4d:z#`ݝL+$nlvfQ|n͆=-/)VF VrQ:Tm-FŹATr,zDq>4L5Z`(`4wFm6*GT KIwPg680ݳssSd_zOYS}&I%]8C{J’6I&h.&H]œg ?u 9u K88afO|?S{yZuu?l5~uu/,B=iC ̤~F {mJ]%t }.*aK۪H˹^O7njb强m8.rH=DFU6˦9ɩ9&9Rf=zHwЄK뇏GV#`XAZ{63#o_WY`l(/3˂ՅQ=s>~q_g(/y%W2:?YO(`O0+yaaqZ*8&um^edU)K^¬d༥)np09sH[G‰uX56lz2B=+#s2[TWKZ/ʠBLJKlUBW?컺JXn6oQWL"3RW aEJ>ub*aZW/S])*:|6W3hbw~ϛ^HW&Uj.b3khng9J;di*pRKh$0Dt@0;'V  UCțf"a<}'+"[ K0g|2pӥtLo7:T\L«7fnI^RHV o WjmY˶Zx[-rO {AjV ofU[-V oMǘTiFm O[§-ۖmwmݶn[~-ۖm>oΐ-mݶn[~-ۖmwmֳ|g)߬ ؐg$]3/p=֏$uT!gV.s!,,goe=7xDu9i{}|Gj"-r!b̬!` ~QNLSec8)XfgLbÌ4j>#ap)N $LJ;RmJ)µe߂'ȳ^͕~}q6a2ue\Q]mk;[6ҫ.v ǧݚiңYrkYj49d9m5^`<0"R*"AśI7xnu͚YٴKב HitL*"RafXT6;$$P*  rrǂqat<zEXwa%c8 9AZ rbcF`*ib(4ũsi[OU[n^K.v* AWgV9Qi i(y-GJuaA;rrG m"N5EO]>#R4-ASvP7Iϓ-hC,fFcwg4 ) UbKPKgdf_6ŖW@W!N&bK= #{U1O?gI/6߸dX : 5ƀRšI*d0J ,I{,6?:I/mXh/C8,4 MvI]!Mt]}4&Ե1&3:j!"L^K ."I&Iiٓ}WwxvO<4 dkg 0=:'3&CB'pXt%c_g<]QEŐiWZHp]H @+0Ghrh޺?u#ѺﮒX@a-=|^fv_R@RY*DѶD| $ba7Cb%/Ck~IBBVl} >u DR@6')@ ELZka>)[@! ٤so!()Ñu%siO Ie["HBX@*Z;b:;[p;|B}rZ"h*W՘TI=:Wx9 'Pm h/Ir5 Sƈn\PRTic{VK$BplsA9%GM,9"#ڛ^θU > ÁG Z5Jׯf kვZ C>Sop C.Ƙ,z./ ɛN:: "v1Dى?ٵk%2;%W^O־aYx7eμ٢KޮS~ߍz` })YHtyڑ<5']MORc(8G3{K> SIQ@OpGsfmKÒO ږjUrqp ZLL>2ea#J2Q`{ƥi?jՍվz}z6[1B@L#b^7쥮FIlJ,"Q}UN: t! ۯ'7o 6ݭ$I3Ewq}vV޽}59}&ULdZ,M^eK%c^[@-Dfշ.ND^y\1do~,r8n}a͇~iW'}-U,o^.fA X,PJh3{҃O*Y4:HZJJHjGy, |+25=}Ky}yHj <sG\-=% s #zruzdⅅbʞ7,=/?4L+Xz[TSgְ&m'+֪MܴD5-%;m=c]hNF+ -g 6+CQ1p8m8i}r9vפQxcR&-Iڰē1e2 dTA#7zkZ 0sdr+L^7^/ukzi:<jcJ%·!5-ݧf5!g>%M[RT}N8ܿŚ^t^kW"wrvs"pс9$Qb88y[@U:iV:i6\UQ!tBT#Z9CfO-2f]V@hR1`b .ibnګYɴ`X('S[AEYnTZ9LNZᨱڌ"&=1͓}֛Cq[ɮnkiM6%b:OU\u>g$'o<%7&+.t+5Zj-gbۜ {-Bԥ5׆>Yb@(3D\Ҩ9SEْRQ88 $L(SɕPXI֌b|Zr]،3vՅnԅJ.)v{Wo{v ʧt{1]gt;F9:$>[(LPdHBա*H%S,VޥX]&DNPP1J%Tƒh0YL'/Fo "ȺMZYctϹnCn#];TɖJFC]؃EEVXjF06sEmP:ĦhcECfdulYפ,,xC$lChYN*ơhfF55׶^j]eE9W06P]HY(uȎPm,zBg |1EY)Pf X{ dcY#~zSG֋Eb͸d'? R/.7Q/xuDw%`ɶ cL$!/( !PS^<^>lCha7}xalaTo#7F_E|UqHJZ~Wo, O\P0 # g]`j-hEk!pGZO(n6$J"$֡+R* BC*+*i@P;Kg" }WEh.{S JS1a@z0ɨ""gl^\7 Tvy¿/bLc ۅ:zݒƐg_.>h^z|1#^=^BwXSͦ& PE|ucB4S1&Br~ h@k~ޭp]vBQe`M.Hiu5X|H%YeX$XNTTvAn|2ZDȔLˊɌE/g2%{6j :=ߌ-JYN*z{mؾrjZ%:)|:nByMEhWZsm¢No/P5 m.3gQ+!@%̌Ks4:effı$E .3- i*:W'h咽(Xv1 >F`Nk*gATxVȂKB"Y4S:cbV"xXΚr;E5Z[,mejRƠɄ"ѰT-N$r*DPl?RZ"R>~/?W[#?3Ɛ%r9Yӑat2^, c]+XVF j'~Rxy5v`yB Z*$.)a|ܴؓ#)Qq4Q׌'v\71Pvoc! tڈB%RTfJ"A]pLFL ]@Q8J[v7]mz0曱Єo ΁7aT2R-{(!h(\J$k$Gh'K06:8dWh:YVJbuy8,Lc'jCt$jɢ a,69R1Ҙ&[&zrxC]Dz>- *R :#:D;5#_%Os]nW=Ov`@(v}BV)X \xK"J@z#lR<y/q 19i19/|-JT"L:MY8aVlQH! @XBk?f^ڗfUlnM^tZtY>߰]_v8H֎Kt4[}RHV"Ct=I 䏕y ?UC~z ok]WqΛ3_^O>E0>8\3ȕ~"E1߲\za/g{@'=z [w:oMZ̡ yS:?![},cO^oNa>׏'6gG_Nfq~{2[M3#_v2zwJ褰;c kO8an<|}7Yz3zyr2{|z=3>8{lQ10:Co4r/w\_l\* *!,,κj WD6+ņچ6։/{b\LZ*3ZWzK`S~F 5*'9a&fP,#D2DL*VЋ0 o;Ie4}cEUC^n.V1P샠v )U` K"/m=n%c2̮^|22)M)hk& gYhz3 )vF0 _]mo9+;mf`>e{O{__dI#)q2WlmDْ;Vwz)XbDM?6s+*`hugWu|x;z+j^ReIW7׽ycCh6.rW_`[kM/O4X̎y,?|=c-cz#e5caWvreAj>xSOT]UόVm+g"myw51Pv@g߇qDYIP&shr*gb<PU;`?FB!0./zt•0B 77JÎҌaQ2A#; 8 Q*5"VT-DBх(|zb Ӡ(5,hj,8&^ՂJE5>E4sfgq:{r.Ї.t!oϟˆؚlYN:tw)Ǩx1'y}'wa } œUp]x]SIet=\fo+Oop: d)ξT*F#i]9C"e谂QMnpR)|G1qQ;G!hXN ql\d틻znjvGC%)(JHENGccҡFԃ IR11I}"S91Q7k  2b\$N)}SKPixUjd$ɷ=Ϣ'wۓMeگ'u=osoUdh7|׳N5*ږ'hs<[v @?!Gۏ ѣ#F!L!Q1 qhJc(L ֛(1 !9Gv05EBO̴IcƠSb[,lB*َJ1,,bp-Qnsf-g-.Jp5}o3?;qCeM.s[ }6H0Q"-*@ P  BRL%``ElywqaK%8 A1d c;Mb(4|L+]x!池v18=j 1ǸKFSRy "a <Gى`"a;wL@c p4QtbK1q@#ȃHVB:Yfy)jG#u o-B{ױerRGb!Ġ lJ*Nҩ鰭10u6j/W %o|̂^ }I{On|.{}^QIu˔ծ%4;,1/"j)Q$Ϭ֪h .:R[n1˭y~vD vvPSR4@b A1H& IV,I+ƙ4"G!o5U%`騔!\zϒ@)t1*-X Y1q:YmW3u1EaK\M`q=}_yCr:e<5IP&P_u~aFF&pLkN+c>PBACcSbTχ}Ik= !W h/Ԑ)ޑ[)K$P#7AlyB =pYM^)q3/\)vMGmI@ IOΥ#W#x \G(SDrA>O#R@MEb*rl<|QJcG=!y`,$OIS@4I$( 7r&cGU;jw^s=N~e=o[U }-y52>URe$3FO6Y JJT #++"@պ Bo`+'A Oސs95S./+e XJ8&'s* B2dH`g!-Irlפ 1PKwQ߼3&H|2 Ywl␥|)YB ɋϼwyhi͚6K\u_^NŲ[9Ǟ;Ҕu0vtgP.6~Z}+@b_+M$sf+I S[4"'G/KREOӖ;뙍,2ǒVM {94ge)C$.}b"!ƀfgBL l"HB2eD"T'r™*1 1e%>i48h}jqyTR.oUG ;z&ufliysJC!M!\A!T UZʶ-9NjSzqcܺ"v*.r hp1s/3뤞kAi64ߌZl_{ql/ucъ!o[՝;ض;9B^#|ſWw)qNfZt#ui*[e5CL_&2WF•0.B@`X*9ܡ0T"duTyϦݫwwixu D`Q={kQW߇ѱL녘B 0 j> K$.0W%Y(ךY +z:HB(gNehHOzZVӳG=ȓ7Q-T/DTr.|k 䴎&X*V%X1fSdROSZyR[ZQjuǤ(KKGa {B6QI2]P*ؤRSEnU6L451|sD Dy &g Hd.W88nk=[O<|]vy~^S[>4H>wl2>/ppJ rwA]?ޡ&~ŸUGMAꝦ}d  4)?2$kVK} 6cJH?Q TY>Gsf ϖɈ\EY,ɁBp&R\7]MEdž⇫O9y>^rԏLd1>?/_WM!1b˫(^r9g%V^m,vrG>xFսIm?}b䗣O~?p`3:sɴM>V볽$ڛL83]QY j%R+WxJǫ1j.sY~aMyz2'>zUL.=^sr<9_9^V*U\ꪽ2VK02,֩/fyUqw2_w'vͤqMO̎>/?X?Ï|IK? XWAsxM3Wx|,Z͗>@l+M(FPYfux66 B[jd6W}tv0>C99>:d0GTH5 :TxRܹC'f'ʖ*Bu"(j蠆rz ЄHOް#JAkm@W.ZٚbzZߊظ[{/ =*]|&̳󚪘'_7z`ˑgtиʦEXfe9A1G>¢lW6`|7W+t}W_x4|T#y~ l+9;?L{ӐެY"/$d?h ^|^ߖ@-Dfw]y>Gbs>0dl'@1$fs}З\ӶM^+en:$݀EHRpL%kD&%% zTRTPC[kQ*;&)+?ܤ]ᬉm=qI%.g8-bp':{dv w!>n^qh!Nn<;6ֶ->2Hq:d]q@Q*id<&mGkTl&4_Yh) S|3Un D k_վ'e`٦Uww0-jox9NF]v΀A8c6VKP k*1adAKo ʣBoM<38&.n,{}3ԭE o6_J(2I@R$!C0]5]*FFMȉN.hY Ng_{*+JXM ȨHYXzABAB>$-A1EZ[JᓊsG볣:cAbG-z=`v2(dKFȚi:ZN4hjY@{'!(嬼5Y !d`* N5(bmB蜧 Yֵ:9"BWlhhVsI 00QZ`(YPJs&gsߛgA={ZD EDN{k`(1PT&Y&z(Dك An鴣l>=Q'*D'cぴp284̶"Y #EOI%c , T*@d:o`T̞S1{*fO쩘=9dgdKUL6oMbQ6F0rLƣѦCY1!Y;`uXo6gZy]/ɃԵny/W6N2eN^oߘWkWɋW,{2~S/}*&g&u(,D P !t z-9!l*)`ZM(Q%r*fLbd'J2mdeH"hB5m++4N#r6Hc,Z:IZ'Qd(P@&d%S8[HW2[p j zZJԤ=FB)ǡJ ?F~"KiEI% |y-`[g e &﬍$ԁș0!&T)"逅`ZV3U+u;IWĻW|1F]J%JkY:*/$,)ya4j%%cHItфq^3xMOv=e_ެo#39U &MhD%RTfաJ"A(1$bK#°]ƒ 5C]#-#&8o0QITBHȎHHюU_Ut`SgYi\0Yi-LǫAՆDf/·ܬvuzP<ʃ}"'S FwIDifD&ϸ˼88CduOZHӒ?=Y(TВ7e Cf䬐Bڤ \ŭ 8c;޻#<$/O |JJ q@Gτ( ]{)۷^8{lr1Z~PXG9.*: B: ylUHbP1$2ʂJk\J.+;ƅpΪ9+։XY(C^6>zjL3>; *Krt6fP,k Zzm"34鍆%aW<9oZo\C׬[`<}ca[6?Mku5Ok`蘒5]]:+Ș.`<1Z*eŖ1:K!Z>cJk#kܹǜI%EZYEO2t)"^5{tR 2-mݶ8;zGEJyY>^<<~θ]]2@%'}ݛ^cn|~d۽O9Ocrck-w5/_QcVKȬkJ_X)܃fvۭP.Z>1&Q:m]*ՇͫC' }7&F/TcMZڻ tGE%Zr@%ʙ= Z˔ږp9\% >gTbuYwhB(1xl!lFΊ2TٛYRfqFCQȁ4(5m]V *4LgX1EZHW"K)iF,'6P?Q2A:8BL:H9+ğǰ4pzQQd܇h[12U2K~ݝyϏ4.oޛ8O)u}jk 1cEso,s}Y0BC .Ր$iKIFڞ̩NJz-"UO8*>J \&?Tk،5c;L6uu!҅ 곝Wdd{zwhr1I'ڠ &g_ƶ1jZؠ#ABYX)tq-ŲK:0kyH TMe]:ZMNzSl E2ےrac݌5v<sSv3]v.LMa V,pfM0N 竅(e6\.4[Ȍ ATvk(de,GQ|!D*Db:ly3rׇQ?9qS4b3XW#qЈwqV;V5vB,hcR.,]&d?TA4hXiY)Qg-81b[R;TkfUG֋yUMu6㒵V#~ɮA/~+zqAT=Z W_ ^XkCL2ۂҸ,$ dC/nLrVܱ>Ӈ);-ܯ+m/4\!n~|G}crw0aLݿTaG c Sl/.]v% F' TOl#:"bkc>nؠxGޡ"(`IeeP]Z\[]XF L 0{[ JUGAuO*t+rV4˛t"zM]y ܾ wyzHf kVLs_N&Ww7z`Gy>^ڧ ZJ6 k69s)]6d.S6Q^z'or~N. A ^hKH چ^1vuҠ~199vC_5ER)+S( un.LIHcVꢊ!^ʫF[z;&ߚl/u\LnmavFh?|+R .x,n<>3B|4 ;lZ?y4j4G;Ϳ:_#Ow/aT}6ѵgMW#ٻ~ylvU` 1d`\s^qZ{/FBt$mnh0r0Vt򆅖OOr̃ǣQ{C.uٳSo?f乌ԁ _OS|GeT6lx??T;:vv0}f?:?|{ovA}wg,n F?_W'5Rћ9iUߍIg0.c@݊ԧݑ\7 X[h਷>@utZ ۫V#ζǛP0|F*ꖏdܿנ\*ۺ3Y߀* ©O~u;hr@P 6Ilյ$A'V-QmNbHzjo\t|$\#lx&*S!MȦ>"CdBA!#H4iFUM&:|v~1'ZĦU_#ݺ[;_| ]CnUC3/yP+,u2w;ߎ^)7wK_x|bT|3_y&o O_)^<`|KǨa产_~aԢ\j1R+EO)bNU{9k,^ty3 DzYL _tB]r_gւ1^$tztC]`놙|K_Kh/5;{'S^hZ+|=*4භF̍  uՑ:[`zF=R[uzLMN NK.f: @`Qg y5Yo"8#Ds/*Z@>YYDPZJ|FE$!@,6@l!:Djk$ $4QBNЙ7 ] aM&x LFS,ADtc + ƚeȑk9t? lڡAJ4>/$̮p*`b7Ա{oA7=hMUE KHՅeDj}[,|IL%z /TMSR\ a)c)zo֠ɡJ7Aj+[NҲ⵺Vg&g ;=}ǬWګm-6-nw ϴ{!חP;C A!$t$giɟR@Rt*hɛ PMQ%lTn'~xxq,WT߲H뙐 Hkf6:䰴~%gX;Ǯ,;u%DLR3t5(ٓ ꛀU:] ਉ_Z ΫirY}h"b$sWsvdIp2V'lx:DyxnFdU pD+R6^ FO,>JT.V9 3PmgZ4[S$ >gTbuYwT!%C<뢴yeRaB=m %g=]VջnW;Ipw*;l1?#uǭ篭i뻻Ott;[ۻ/ =_iy7?ݡϼ ϯTx:c h =reO}Gł5?oj4rz_Rls+lq>Zm߸>p;S:K:$u!uVc[(g<>g,ʣ`LdiH2*(OR|Vlo +_рeK5榶޵Ƒ3׿"S|fx1O<[} `Yȗ5H]%m@֥9ݬsb!t%T*z$m*hot ].+ *Yw3gO>W/]ofgVF<l?6KSe54{׋޾c5pWmdq>TAY 9(KjV=1StɃXbx}PP,(SpX V%Zs2ĩjR^GyMM nMqr|\S,f]1)WRn%us幔@|AMfˁ7v *1D=*GU&6&:/|g%SBt1ւYYcT2V\ 䐨 ] 2Flg|(Y&r'DιTuV5Va՝ W7sv(4&u6ύƇ̫=i.y>G&8K_kw?y%Zm'_,^HG}r~@uJ-Ge>1  }bT6Qg87#8MMQL3R]ci-]"e&fTtPΣjMt;>,Pz?yce[AC˗|+^ly1gޤ[/wY<$I1iO)(әәڿn:ﳖuҙJgNߑΜt懾zN \NUV۱U&JZ5J>lq${;<12bgĢm\k@r'-ws;i+m$tWvﭝ~ZqՁC自YD:G.ci(jz:5(Jh[>""M U3ձUYW?!\nm)[eghNYg[b&ΩbqA%\VdY"()R-h\A13R^ԅ -ʅN>ki׆2&X%!/O>{Qwӫ_?._\~ذl3{# [ :|Erm.nZyyt333ވ\_u`f;>9ޥɷ. *;걭s=F^1l;4)|\TGYGعY4 Sdӻ2\Y~B^vp>ȍFpu!:(QCF8 nؖ܄:l"BI>SMQ.ǕhQl$99*GGR WL(ȥU=Z.;N.=Nu]а;h\tKɱo)聨*go+.Pcy+koXN}xeBXGʻc2rI2 wĥg#> E !HɂԚ8 [zk-.=&s(J,)T͜؏tn=P,tZccr@U`痟?8bv.Uِ/V)0٬B * T6rp+chb/d0F& ECj/dJrj1Tٍ"I̱v1vjO v[eMZ#` :Tl RN¡uƆBf$*:a5dt.!Qs嬒bCTٍ Q"ۀq,ح3" 'Du\Qhp ɹJ00 ]%Qs(N~u%ƅ WcXҵLPਘbbq)d1Xq;#b7sgqZUGr1uv%"U'.ޞo 7?+l] @m(1&ĤlRS5 G˳W83ax v!'a1FUx#N;-(1{MگMb(\8ћŻIv_CcliCeh3 E*0;>G]=ƚc>oQEcېA&#WFB+%,Ns,4MT`kR|p |W51S LtSڛ@2gtK\S$?\ps2/=}M\qҗf~8x-^vz,w_e'w|aG.Cőm5ju?㶜s=oin;v45GBl2AtQ'9 dKP]58ajrӴ-1QU3e_y٩9^ ;N?u9oFn5mp~vn!_69g[azq^wR{[r7v}pow|Z`3f[wstv;NCpyq?sY-kn4ݷDΆڼH-N"Um毤-W/]ofgVF<.tf{Bwi|Cx}l_~f^.Ǫk-Gtv\&>FPc -lPL7SdǷ6̴VkkG_.~sIf2ؙA0 >F:ݞV;+JnA^(D_4"Fm8F2M 㔦Jbc15s 3+|Fºz3S&y$tViR* | \K-x8z|߇zgMx;(I2uu-n[ҭ ǧG*9L*qYIJ*#Y|2G 4  ՒFYʫ H]tI%] ^T) 5:ômTxGDJNcRY* fXˆ l4u;!}cPig&#yx8̃`]+J,lpsF0VqhJ" E1Xz.mV'Qr䲶?Դ@XxH;+C]d344qżY#yy mwƣh6{N좧.PC!(j8ѭ>N8_?aR9E5Vƅ?iGkoJb#<A}G󧪣ΐ jE\eqb8J83#_s(xV*5Z1S+)kF {1& DT\s$SJD#+pR` 0EJA)TUb 0e{o(E@r85!%!Jq 'l.~:><:=IF$> Oil|VG>F.Q׳j+t|8^*/xu*Mje_my,,`Tz7;x 1uwӳ›e`vCZ,k0YӷT/v)HTƣ/5Py5C%E;[b|uK_!]Xg3qmS/Y|<^:Z92v:[%Vg/j׻b*ZE0#axRU(ƱHfbܬ1JR:{S`. o`N޽y]w''߽x:y?oN>fALK]"p/QQ977uRFr),OH&"9"XA4NpmݢÛ"R47k*F ֭.;z\ڄ()-Ts;ҍ wńߥd>Wr]E#*;u,$6 ̢`b )!  OM`f $=NcGf֥Ͽv)0;?ʂL$BZMT1G f|ty5'+\^eQ ӋɿٗNl5ioDtF*_zK)f"Ru\iGQ8\UjA*m`T{̣$'Ѥv|G;tnDtn䉒Y-bHJ,w+462-+I'_a7Bj,w'CQC9$) v9tji `lєb'(HD{ŜsaS1&C`tχF^$)0w̆ŵ0DT:ci=SʍUdzc9#SW#` J &9Q0lX9W0΂1CclΌ`kVypu0((k*}?0#Sךulд0- V0#*)dFXrεCb78ݶN/JE?q;cvC!fH Y@SJx->c|@}eSjqq; żC_xe*Q1ri3zaQ\}:7ۨc3ɾjjwmFtof0, ܘh!p3k 5߭nh6#'BZD!i`1|ue$"_jY"Mt=CR@YO!êڮpTEᨎ4C'TkTm.ԙ'NJ@m-M 遃ƴT/p彩>^D\9|O. \_0<>R]LvǷIJ;I SdQ,SYz"Uw 60u&p$p!"lUpom^`e)ypo1Qۆ-T?㊦·yC s3<3Cś_~,M2"+5}aX\z>C\g6G gOXff$)6I-,Le $N.'0LQ"4mwsL(M Ŕ=J#"ӂ;X+\$JCFĹ{u+Ki7ܟJxYfAt 1Q#&Eۛwv%uI0{jOqQuThNqHk_2O]ڳ P!q;wvJ);Zې*0sFX-1HbyP2b(b NBe0IVe6<6rgSքJo j,b>ĔLk㯣%#u7z.t`K]L(bi%uG[ 6_ ey8t#{kD<8;8(WT!kQk8DQb<)ZtKZmgc8q@XAM 3d~ڛ}e}8 2*AI$-w- o13)NNf !r0d9-Q0B*L+\E 6!b 1+r-W,Yyd tTxD@)n %*JH-ŎL9JcxƓϦd,U Y'R7Gr3@#C ${Co `%>[^,O ޟ^;M qĖcG"x̀˟ѣC4"HxiG`07w2)֦N{` -:]B  B'ԇO-t QVc9LI5 k,`:XHw1pJXUQ҈ٌeƯCrv.8La[N\/0dtj"eʋeځ sݵCQuGW<ܛ|2ը-"Ѡ,gGVq;*2~֘U"h HTGOuW5G)y'z\2A%wKk7@\$wKݒ n/0R}EXI:]bJSAܞH eH 6 gc"&On%UiսU=D햹0WvΔw͕rճ&5Hb&iQl{Hsqi[sf.b_UV]7WIJEz#ő4}vR':ofR%)CŮ|Q׮ȓ;"OW~t:6s^pO(ծM#B~zߗ]&̻K۝_l{0}5ު TsXOcMvTzƆ̏@߿:xk.oݓ[uKt+jgڄ sVͥ&Riٳ2AfY=+ge̞ٳ^i+~VT/,>sgnM37}Ї1&842qh&ġ84f~C3qh&͆!fl2qh&ġ84f˓C3qh&ġ84fLC3qh&}ġ1!WR )εgzC̴pyqfF~}*>Q}?wdmBջc&Xn#j_UذvN팯 h;jČUc&[urUkU^Z%RMUL7uZB6VFFٱp.2Znv:7zG[iC,#7O.$) T|7*Rr^QkSE\붢ǰQC{hnuM]J]|XYeY93guΜՙ:sVgY93guΜՙ:sVgYo}XgW{=~pZ'qdNrكK\U9{ ,1C>]PreK]\4 JcF1K)`x産[IfNjwԋN<;ݳZ㻂o/r;1ʼ?]l*l1XT&\."59]0P,5zuYxXx/Ġ|h؜yae܁ogNGoD"VǼlF}hXŹΧ5 ;:JC% xSЙfCK Mf>$cR俠kGz:=Z("F둲! dɹ4ےL*ZަVjSi*JvJw\o4eF-_xq|VZŹ\ז&=_$O~q\ԫ_-.O/'…7 >/kn>ĐSRNll ԴyzMzʌ .e48Af;?:'*銥o3Yw9m7+ Cc9}oȴxLYjx:@2YkˀUJa0 #:gݷ*uozEv_zHc*!d]Kբs+K\:Tr&=Mr s3k@A[cm"Jj:גɗ =ٲ*{rvLZ7\62fTCml[g62Wp)}+}J7"K)9Q)DjB|}"5+"$XoĨ{)"aA屾p&Zw>k~w%{=;g]w>;=R%i7_'(r僉̞Ն''>=t멕( ( `[j 9p}Y )#Bp1j jE|U3;]ڈ|Dlnsr%[3v#atӅ8cW].tE/2^݀z0ӷ]}Xeݓ`ǧVp|38*T#)R6PJZT%U9eER` V'ЌLFR5Mm1ELm U#D*rUw^8>W׷P!f, YvbE(Xa0lGQr Q19ѨĜ5Ÿ/Sgf8k87+!EQr"@u M*bAؾMwhBtbs/j@3BE: ʵma pEW]D/6u-ٍKvҋWE:݋z[ы>ZBJ4Ŗ|cN* b hF2%PM^| /wug}wӇ(ּ.@=zm:b "n~|GSݣǥDG+D}ʈ'y~nnA֮dN֔Pkޤ]s!#$J+:9tu ǔuf(.9[ P )Q(&]))ݍ-+o_>o6kB h#"R͙-ZTSIP6EκV}*BjT~Id.WGj05m%7Nga%[ՙѵ¶_ZՋ*jMr6Zꢬ\]TXf_}XO4m#HU2 PC@NyWSAk8JbQZ;#ia28fR5 .Z"Yd+JCT[|W{x}YE7m5^y5ez/s!Zh 1LWIf[t؆y?.%ɛ|ˋK{jy / 1J^ ϋ _5'; @+ScV[<ȫsQ'|s!(nKVAr:R\2 >wkaW#|5&I/_jF ğ慄Yfp?,QWϛoЋY㗇yb%Zx~:!7Ѽ;_=r!ěV%5jf\yq.^/o;"uts}wo. >!F4c lq;Zoskp-\-'_ ]M1/씚 H=0fj&(BK<`Ŷ.?l&zu=ɨ09QyȩQU۰J=j6[_xqZMm8]ߟ{n'r}zG;Iѫ_ۿ?~WՏ?|'+Οd[O)b=wL9+eo$^Cx--/`q]G]/xB sۈY]i|0wU4'}ۻ=^6q oHdR/K){| UlR+Fl(v(-FS{6e4ž+i 1!WB+ΟK!+t ZDr>Lё87ulUB7+첵ՇV~-\5[>O%(/]YNє g4 5i_)* 쏤I Ip""4UE'9jp/v݁P?vUWF j*)9*TaR`@\e!ۮc.w=%;dz2CDH^zR:DArj&덋@6NgK,'DE?w:g Mwں ݁L&jy)t_/epu9N{oLO-T`-) *`TJYfUw:v!mYt;W=Se๾, aOa뒳V\BTbptF>aI+JHlrIw:˼Gy4;hLw,i ܒv-o03%[RJ˴-YlI2_qlfL@vHvڷ_6V.'.2>ht} lk\B#?Fwc'Z\__ig}i07#hڎ/րe׻} ˖Ė-L j~݁o͸u- mSGs -E VkANR&𢺘ORנr;W%Y%8 Q *5 6@Wǁ^<z;Q ' Gu,z!HX VR}W'?K%u4*XbŘMAJIXzo<8bp|xꈼQSNܕ=IFDO(1*1I JtQj#^eT6L451"s )K$)P5"ēPawb<mt1TyN/75B8zse[>̇ lTׯ JQsfqUٽ'5 %evvOzv~QLgy2Ec_W6ݤ3կzOm%yuϭn؝6͚ edwu=ʷo|0v[^zp}ly4fmnC8.ƧeLob55lכc;YlS={ZꬬNO`I~rl͹fMfg[ݵr;p{a$T2P BUky}ݏvĠ͍mJсYn#x9Hhhj-UL0ǃk/DLjM<(2#,"@'DLa۽;=mCy>ywDrW-_~ CzhPˋ\OK! %,$c&6A38GVXU)N  igRi6r V[\.it*x- Ut=b8 q6SWtګre9n)Sۮ,,%yMayZر Z en`\R`;,\"u6ը+dWS2s+ȱ%A]Cue"u毧ofWkQW[y ͨ+sG/o > vyS2ø})t0.ȌAu%uuS%SOТ*3ͫQWȮ"򵨫-gǮ2@u[tu|'hU[˝XaU0V'ѯ|a4]4M*J+b=~""$gڽ{ ԭ`^^'MQzvaǸT9g07}lQu櫪+'DLux`35F9cV7-Z{*JI>KHE'rL0s*56SDqd&VN|ȼ;+1Yˠ odŴ数U:V? 3[>&'oVީ+,(hrDd5'*%żƹj*W%Pyxnl $e'ǤIʔ˒ xZĜSaJ⚙XabwIBt`"WgyH7 ' *+5ߓ++7IĚEpp$裩md^KfWײt,KÇ 9Z.N#'F{N,2ʂ WJZin$l̞(Ny$]$wKF*x{.`]P 6z)?Ց613ܣp %%Qks.DÙ!I!] :`L"\4䲉Bg P89`](,YA/Wa4v tTCz@$z䝝fr۩5{9ycxUL\Oy>ϟ"O$1#JFєsE%Jǩ)% XgO3_AtA55BK zji(5,hjpLZ]起9QOfNil4N'Zj8#}XaYҿb2C+oe!(u=Y(FiX@i41ԐF?i WOsď?mӑcN?,lP罨>ԛDeѿzE*D/Mx3:qExONul2!:H2\Y+AQ~3w1mcsP.έhQπ>ϐoY?at];i"Li e @Tikh"5\kt|׷kvL{*nx|M}U3TF<ArkaSpu*O LMfm,uB-;:W5iWİM'-Ŭ4ikfsbozV,]x)qL0C6ڞ[Zqi|z{Si8_;DRNX8$GTZ"hD㓰4@ 2R$-PֳJ A҂29#G6]] ɢ0Q'T řfzD]69ٖebgZW,WXo)߹5*UG5|rO^/ʋP=3&[&j)Nz*3Z!lRMV^q@Dv;Y:$ &/l"n~Qv8Hw=[llp4M()o-E2dȹx \e-+q̥ugD 1PKwQ ygLDI`Q"nЮ;=2nV=\=P ˾m\sfvsƸ>{qp-Pq۽『^)IӡA9mu5+RRGZi"E+Ǚ$%L_@^iTs? L[g6KZi= 彊Rw@hgU$ u|vdEC'12+|4zsc/2\KUJ0:yK4atrdu^鸖zb-"-z{#M6fpQkeYCkV|;7/dh:beߴ~ϝ{-@#_u &X󇜰ve®W7cؐ'g3_ʩ/QJŸovYHq\))Ev5rV'tѣ.^Yzg>|j̣?~Q/r+>k$mnRumNL5IU kp%uCnӨM.X>]5?-N.O::/=Q/q,ish*1rb5'ÿzaOoMr#ުl^HuPu2`iHT`NE=?wͺ(EwMޮ->^ h^?3,`ٻ޶ndWٻ@@pћvbۢhz_60j %WӸ9lŦc99[9 gJysN3jIo.uKt1m;|>f*A.2zUwv9aqnƑ9?? X 6 PvHu U D}tOZFeVȒF:fK蘲V93cVE0`VJBݤj&%} =M)z>)b&Mp;+U^KΕ4cZ[E>bU٣\#<+c*J#(I"d]B{ZIeJeÔWŧR2c9ڃ&6P(o寘$ %%ѣ7aSY˴t4+j.T[' ZJ}:]OGD"'=ɋ ޙem #;_.-ǩ=;O6b.3ƍT)UZq#$㙄L@-!9(\{g1{ϊ[#]Iz]t&Eb?J&IHՖȹ[2*daq_Y,|VpEQeA[r x<,oB Z!\&ädډ`"Lh-Af5 kM]U-PdIP=9s-pK`t;6s"ZZC,,sIP1EjW["=mœޥ 9JI"1 /AV`a"c,J8r˲nqI2!CE ɚX\ܛXV}!Tp?y"]v)?E\Dޏ\1]Kvq2&&}YG_XeGKS,5ް zw11ߑ푿#  Zk\&(AscHB!WID&*4I1utvFrCA<%I[%MZ9;>pM|6AۯixK2mX`ύ0O:;8Z3z5ʳ bzDxU :^]2ZaJb+J$ AՌAM'΃XWQXo&i6i[q3j>tިr;8X:ŭ3 tZJZA8;қ(+Y9SƀK 8gQHC'GJ$+J:&Gc!Gbmsu5rv]Z|i'-zK|?-;jg3/N"0)zKj5*I!A2gYB$7:=hr^| P 1p7FAn0TfM48H4# OvAğ& CT܊/fgk̺/]68̕/'{ĨG^qu UiXq ~?_MB?ִ #eБG##:NdJzbVx<QFkD >R EYv̺W[lMr110+6"fQx TnvJ:*gD]x[e]H?.~V6=X1 4w0\}k>jR!`WYvpT K7Gz=8uHom,Ydm G51%_Ʊ & >P$k^ C@Raq` ?/WҎ7Jܺm$(`~۝&VUsFa%۩Nwy(w0c@9Xt$Cփ O :¢b0-M/]koٵ@Cjz vW~a z{g!ߔJlT&ل/'qO{~tM|Z Hxiܞd"N;2ku A)1S.(\@zN!=d!f{%s[ҥi}8,%žGRx\ry˨T>BJKs2J+(IC*:tqy77jjuֿ< .݀z_R[8Kk;Khpy%&Fat䀍O24 ʆ{ȣ ܣIGfpuR'-_'O6{L2c$z^[p&h=ZHG6G'%XγEi?t6XN8oU;|' h!$mH8:.!X9qMW˒9,u'={^-dm7zs6Ƒ;|NמcJ;Ihy KZYf!ꆔpf1p_Tn ig`⸞C9-,|"Dom A:ZĝGgev˲R g{ucLzf#'BZ-+UB\&n. nF,7bҦ] /SR/䕛4#zs:ήFN/S≷tj0ri~Pkǥv F%l&N-e BT yYG\#g;_G:fNA KA٧_%:< &]Tɶ*g7ɫifSzݢ iQ6?ZF\v|QApsq9;{D ^c<:n2`?:<j\M/g+Rgwre[O.,9JiA6q-"0o`F& rҩ^G2t 1\C)|`4Uh]}s?3oVkO*&53!U|q^5GL6ԔXTLj8w9jrK<<?aGf/"\G/EE-pXBRV2BOsш5Gc˘u@ԽeJӤU#ؔ>>=2,Fk2ep\W>FΎT{ZKO ym(fhz t4k"o%m^B,͹+@Koq~Hdп  n_j7]ۛhtUp#l R͆j.SmQw7Fl:eR1wʁ~=pt G'QZ/|vsT^s/>}|'NjnnD9+~piGJ4-{w*lgZ}b}ۻ 1~w V>-/h\ eL9ȍ#40Fv .vCaB\5DpJBPܦ<_ssXB("! XH 7T쟡%(,*tJyf;Y(a#g)[  ehR}j{F,S>&EMBr[͝hLйleCE5rv4T[6Tm7~ ]YBliWgCi0OnvR }̳S-+ڨCjfCFL92VY|9X J;:pIɪcN\-1U&U;(@-rqVZ=[uNoޮWjѤl|۪x?hҺMB?V9B}'L^TTb82&52>qILd^$X3(KB) a5rES[%/p:pb><,#nÛ*2rwmMq2SJ9bxgw6begOlĉp CןT74]~PUJ2S_#As4]aF F VݼNMrA5-|w<Y΄ 2βȲ/&U։圝Y{* k.YP53`X7E3`6LE $T?ű :Y7r6VV0K1 jE [VvOF\BtUsZ^*Xt _( ?jQ2rΥĆ`dJajLzHΈDdT9hAtv8Iu^=R"/QhhɆEF*$\J1g5).cTu7^g=7٦  G*2YUSV(_=AVYז0dC%%-ЕM'H5N#"lee="|Xdl·I謋>X*!i*F9:PBV˂VsF>''`2n эC,=:AmcMp}<2,LZ-D)Z$LD/@$jAÓI}a~U@֣@>0 [O"L)7v-b'#yTɰu3~QU `[2$]R ˀ nd*̤B'+< n+1mw~O*soǁ/"g Դ&cy6pzHQb`!n:z!tg~2ttyuxo8*D```2N~+&p]5C_gbH&SL4GH:8\JֺE =+ؔ3RUrTk7f90X &T% $ֺ-Pu#ap$:;@/C㩨 *R9*s=72TX9 %;\3T]2hckjcF`+:j $]}#lFÚ0<ɝs87P +|;7/d߰GrGlY9BjF#GSGy##f\^4a Z)h&HՐ Mt'7HKz=TL54 x. (@ZV"bRϘtK!6_fJ[AS>So{ҍ<]晌is5dÍcislM v2}˴9=!/E8\_|899J!@R<`lGn; O.!Jf\5"K -i\^y@sJR-UU<8*3%N\G4)ҧwK5*{4g=O0T5;e_jh hPP*(\Z:j1pԹuAQFmEO22NK0IQ `FrtNqSً eO'K5n|e uN)#7bj}}g aV~8&#$`JjOV>rE u .EӜ5RH} [yky-Х`\}49̵bp%AК0TQR5c7r֌J7]؍3Յ0g߸ƶĜ*l \VYkQAUYcu-Y% jl`(Ќ 6d1ZxSb)mM丳F;5s,ZwEk'{omqVgMb*,BQQ9ӊ\'K`SP9 R)*bW}Xɺ"Fj+:AqkrQɚb8YBRVIBRyX.UlScшc[;kD?iI#WTcfDb&GRպJ*㢡PX}&_g7.J/zr =[].xe 1]4tBjOz)hjymv TeȟghGnTUx௹"b )ݣ[lsD{wZCcleCehCTxǧwpx̧wT#wTwZ@]ZA;Fƀ+%(@l@&HBȹv8c֔|\:h=bƀ'`Mjo݋ J- 8l<;:YakUSPqPI2uXaFmNMLb&Bxu4߻}-в:! UF ~i;[{nD(`V^83x&DQs6ϦuL,[[6j&mC ZN,+ s F4Eh, Cw0m4<9=yLpn_LvӃD|9d`r ݬ/Ve9!XM TL}r((pdJrkV=y$B{:1x[bQ |b VNk 9c(Y+ucӍ%N Y]ޫۄ|nby"ޮ,,'(&bsX{i M_ s"` tqSZY0r+ ج,uޗa(J?m'^zw}<|llWEjZc &ε@8XD{ yReD=&عiq`ԢQ!DIdTl.m?%k*FΆ%{-y-7 U M+ͫb 6Y$#-u̖K `pK5Bx@˅![[ ui.d-9бqC1`mUg#+A$}(Zy\0T;K*,pp$7u,갥C5[h5Bdb->X yTS-4=?ϵ5m"HUlrYZ0+nH1ɛ2f E+tRodAP)[yuvE/fH< Wg"Yk.҇h:? .MGFLؓ.?5[RWY kü?yfY䇲sjOͮAyjv,x~p^ P-Cn=jC9J2'jwY i7(~ޝ<-~*m+KUICiE%I1.͟eһ6Y#_⯗kK B/t{睓P"'/|* 3f1_P;M=%C E˩exrO]67םsxzgxs1ʟ Wݿw~ZŅׁĈf-§Y=<\[Wk{I3;|~(Vrq/Ժ֎tHG1놱Ja3/"̎IVbŏgW =ldT^;QQUry;(+iߎOSy9;Mm]s<?_~){ Of;佰D_wǿzÏӆ߽~q$+pVRG caJ*fATƋuO?_e^"-Qĕ>]ݕvH z'>" -wW]ebA߈VCC9 6d\95x=P!l~m1K=,u>ITWfn._AWl$l|Aitt1`\rAc4b8D8l~2۰.B|aF^PM1SlR?,\SŒړ!O:]ud45>zĖs=m9}D<tk-\5/L7:6&ëZ: ^;6*:~VRӄ[^/[sS/&@ÀF Κ3Q `/͏(wJsuTCB^WCQClMˬ9 BBT}^*)"6BmD:ej.YNQ1ZjFd{"gnޠÃy kSHJ&A# RUfl\FJ; L%9nydHB0ɧ,I%՘J9`uDDFn4c;m:*H,Tq Уl\咊%+6,a]:ʭj ;KI޺嬴uդMtd8ԁeXQ KΎ&e`Ƅ^@־*@[w;Ӌ9p_Wr^J>3Y4[][헳S'E Qއm*uscɸWgH&A64y=.kaOwm#I{;@>ffb; W[EHr`U-٢DIt, #fuuUuU=ick U cJ#68R$#Z{1[,} *jZﱒ!\#q}r^ɍo6nT^Q1qi˼*}(V \iAíh|;jHˊ,veE0f0_1,۞ \X+oF>p3Aw'_K[Z̭5m4u͖3YZTqj\#`9VΧaKٖR2|%gS4ag0zrVֳ7fՋ0.yvԌQ-[i6ɤJh(հY zg.?a,Y$ Ed(X 2@1꘍t>7?Ҧ2RX f'X,! bRzUA(\gLG ?LwZ D佖豉hjs:Oot<m@c8u?+ɔl-];O,VU$Ch3Dd]0e=z3yl ѻ1[)٬F.@һ4(N)@yifyH}-nN=.2?n #WnڞM9߸Fă3*+5(gQbN*{ ^Wp#R*E~\+Y+ےÌѓiK'Ӗ4q%M\jے~3mIeW! uڍ]şW]-~G]%ǥJfuLjRLOH]ÌQW]OE]%n˺ڍDY]uEWR59*NƺJ>zJ\*KTW)^@n˭h2\6\VEf)VW'̚X)}/ ޾P9|X\vACDb-ϝ ʛ`X=(ĽwXi)ތH0bn("u,@?+^zsqϻ.݄)U.g12rIu˜هA۴j\&dܬoTw?~cgmI ]/tJÃ&^2%(7K`RGTRJmɌ6>@DcrDdLĮBb  c7fS  +`XJ*q* J\.OY]uUbXQWݯ[>И<^@㫪hrfΒXŊW [}V>]6W3rN$QTR?pU&uO?Y$KWׯ~;g<)_Onj{MEyZaT|?fng/8*e_;a|OVZÃۿZiRvFxUCuZ5?c~-1͊$| 1jQm O]˟}Ϻ\V;-Ee?Cgh -3|Zv;>t3Cgh -3|L%4Cgc3Cgh -3|FPZ>Cgh2| Cgh -3|<Z>Cgh -3|Z>Cgh -3|Z>C|oz6uLlJ*(ܶ4?, a/ލE4u雿 Dh0\7)O`ƶ\z/<]T3&F(qyjkq 4<X]_?.f |=L"S⿽y83,~t,y fe=Jns^)ԏN5M7x;S?8xMA8뛂T/0v(BWg ) ^aRF\4o+il_ЌpG3w\_a& ^xzu^/T,_-/VkRu (Z|u7?-/zǓy$VATSx39@5L4A^ ^ڎB&GQ CVutt<00M8%.taXOZu}LV`y0Jʋk.ރ71o|Ud^m"ߢ:LſoӏU$˛/^.~ONJiXiA@yX&=#zN"nO~qԽ>Y'4Bv(@U;<;c#&(f_f7?׳Ԍ)oFy:&~\籁& : Ͷ\tOtiy= 6%_J?nfv덷oC1Y-&jVH١FxpsÊβo~)Z#<[qpuyع"J[-2 vqĂv 0{$D`Jc%Ύ) C{C P*Y*EǸ.qC S X2tie$XLjAAnoJYڹgL`:68>Sw#6A0ɵ r B($<3`MdVbaJ[=]jWw\ݵEz^^%4pnTl(cWD^Q1qUu*(V \>CϤh|;jqRZ%n -K5Jޤf>mrDVI_EGpg,ha3]/W߹w"֢x3mu;|L3e )ֻfian-r2=vR_{UɬeJv1F9˦Ky{|4!#z,)@hHSHȇ(xĆ#SM%V+ShƈsVLp_ MP7k5f,`ZF &6MVHKDG1ӱF}L6zn02LQ#=_b: 454h"4(yj$|XXkTD`gC 50$5AZX,`DΊ|0aE" T)poDP{cgB UYYL?3˱&'ʞOlGPULƃ6U$j.͹ *lV[S?+Jl!-: VZC`#ۅZRުK"N Xy]r7}<$+Ǝ=aKe>IXNTV?9MXխȘL; BE=zsr+Ñ~mv#6sG}\enb>c7-Zbp.y>p//S{/_\#vլIg3喬89@PrgJȅZqJdaR1)3{d ڎicj9#vߐOESAm}Fg{WCwQlΒ)˄4SdZIf"{LɥYW f*(d9AQxudգs?EOcnFψxgWP1ǠB2ɞwDlҝ|ң!j ʅǺ\@ ǹp{GRm+̌P!.AJ m6s1dƽ#//&,?; }a'Ԕ?y/|řLÎ34]w۞zbJ3X Z4$",쐉`5JФՌ&2t+ e0Ӗ ~γ!jxLd+RlQiz.S^bwY"mk5-b#(ÚXR*l :FќHjڜgy`_4 @]@\zPM,$FgT5stA0CR1Q CM#uOzN)?K_p+? Qk3cISQP\p.!W6eo_>g>yZEČ'a,1Qe7RnBmYwgOm1i|>?L%&d6{3 *\6ra@]52smތI~2g귅SKy9זe4 cis^PӨpzM%G={RlfI.Vax8r:}o`tgH'7|&[jZ+};MK4-tXi_i fMK ?}$,N"?6Хd0Sm'fk;3@V iVrc֣=>kg?8@-P٨~|=ĚxbYTI%I;h[KwJuPha> GzI=M^#˜cy١ZP7Gϧ{}DnC8>?HcvTӤwEv3=Zðs^__T贇М R^˭#pSii\abdWFܘ{݇#JÄ/?޴ ͢Vry7M֓}@O_|v~i8_)VԡO{Q="Xɝ;Zޘ ]4q='{Q|@ۦ-3vOGi`ݎxfn9§8E\Vʮs"3p 71r1/z-NR#dM$x!9QH9ZR=@^q"~ZO>/?Eh݋}vm.]4D5L:{\&)zѴn I 5rc_#_{t$K C {AֻGgf0vA$&<ُ Pxl\$G@PŘ`) giM **h^tpZ@ׯ^(112B dXo\enhB^֖jt Ɍa1hŜb7[w  zk^ `{R# ?xjD뾇gπ %M(qL-c y{L(Y@#$5S˴:KY0tjQ .a=^'TH~}e#mI“NˡkGwdMHhb\`p<4!Ph9z6f 9(%ւci4szL| y/1T݉ńw[ͣiMGkڐ((E1"SÀLvdmX4r 4Hu$O\/R; ֞#J-!$+9| 9uђYR&/MiRTZIsmEK.pV 3DTGBE@u&v ZA-M1P&yDL;Q[$`G: \ y)O4rnjC%rTr>\{J|d, .)޽![R֏E")V4qWƒ1msUNc'܂*6ӺL+\d2ڐ8V!A&s.u;n56$ԴѧQL& ܃3j3t! 0`Ɛ)~XoEa D"`P]g}AZC<3X$ u~{Xjcp]Y8&܄!leŬ%;V\ں5XGe0Ƚl2:Cpͮ wv!D\u;& zvCg[Mim&[%-̷Jq;Njl E=a`oٺ+k{#.ɩ)-ںj5uENa]Lȿ* {a3baw&^8w_zZŏZ)!EҴ:/jZZ$ >>Ȁ:p1Pa`{kVi (S[/UYAm1ɣ.&cc@|Vc!;"g<4ED&j3*o4dCI=U걱\0F{q` m@^ZzlZJ v֭(De6 eॻ+[,GՏ4M㨈UX5f ]6$qDGO7_W;_)ꤍuv :#+xmЈ2=2֫?CHb΁6DbxDԮ gOS&t ((z(۽XGakrk#(Z⎎@P蠼ڡ bAĊw#cy63JCX紮T evhé D֤ x̫U\XT  38~暭6P?z`bvPD%:4K򈺕0cG8giJYNX&`kO7Q w VcqFS (3kx);/1 ֭AvQ.!XPʨ[%tJ5~>ߖ4U%a!Jr~{0.^oNeڋ17/e$~ ..n2%,7J6jA؈/ 5;Mg7.(u< Y\kMky (Ys]h7h3 Lel;>0"I*gm=*X!2$phJ @/OFSԃ0VV%WؖM rVԕ"VGP1[%$\J ʭ,2h #H5%#1 Ui,%|Qќ1ΙnJNɐHZ Xk&dɒH:B,e9C?S_])3jkԶEuL'xCJK nM0wQ!˜Bڙ3^lq ySQ;!j|dN [gגCz&FAXm(\I2xp4U֘m޺%`U|.պLD!պ:v*6)MP$1tAbEsk*o#7| .-00MH2X6ϱ`Yr$y9~VaKe=5@2c"dtT|CWra{'Q4(B- L5nCJ<^x("eGR *<堦.5x.yP-0q޾]~̸j O-W=d!6$.8̑.ҸRW% 2&;:L uBEr s0 _ʈYuYF'ecR`.UM 赏giT͠\I+B,)-es}뫟h%hsEۇݖd&~{69&_N fw=`QމgS?IgMq?j~5]{ow$]-R7[̸~&KS*SEd\M5H+g\bKȸDjI-ZSy>?Iio,0T32R$g1izK?'cY {%ԿMNݕ|m]crKe`/?[3edP^;r$SEBVLaU`Q*)SuO~Uw[ӼYّO<[rRoVeՓUm&~6_ |ߞ= ki#7\R/b<.ZG%}{}s.d:KJyQu _r KVѦpfEKf1࿐O"C k(<)]K^=thzG2*Á7 d %v_mݻQݕ5}wL=}y95`ݿ >Bl>v5ϚCkw5;Y㿤e~ʲl}fdlT!H3R:µd),r$ZiBNgpĹ{oy\Ã\}v =<>ܕWgLG I?"HD#G?"HD#G?"HD#G?"HD#G?"HD#G?"HD#_*Qs1K=#AZC9' /hs3#~,OBIp5#IOa\Hoc୲vS$ߓ#'aZ}ΪQ3O۱rR:KNvEz(O3[;+aSewD[μ:ɬe)! %I(I^G!4";M",4?q0֌" ԰BOlmo\j}kVwY> l'Eqp.QLrjRim|ʌv9P0pߧQ[k8Y&>ßf-D \LAz#e'd**N=P M:Jq?~_jig"e"GbJqc%WQ"M浽^hⲿh. Į*M.x4)z` #6I\:qG̨lLwյgjKWD~l~CwWo a@3Z`gvշkDhuu?i%yuu?)<,ue zjcl9uU]\]i"ZX9@_Tknl\wJr 庽.zW-o<0Vv'WMO7Vl^4Boߑ555U% BOR疒Wp~S^3ܫE?ܝ ЌO+wz0@M׶Gi#o |:~ssW|Fz!rzƃ z0贴g ]bQ7Pjz>ڈXkYۙN[{J$ NTZbk/3+)y56'2~DVF裱28`ei>t+HvLV˱2 GJ*Zy dTzJK<"uUhUxUְCWWEJ[/R]qIŬIAl).dZ%|d ^]Xn^իg9&@lLdCyFZpk|h *`O{#97̮y~nCZf֢y_Xʅ ,0Q,9[n|ts3tߕPޕYzzcI>fg՛4_ՅJ{g=ι74\Tzfd˸5k6vcWlrOzm&wI^oNA$q,ir2=qUמ-]͋?{WƑ|x}_,Fl7F7>YS$WC!GRK1`]3~t~Cu̮X_x5p ̮ QiqaTثgcKخvxrp{C'E{b|sOMݐnĜ ևO0i*>.z|guQMzUC6 BGf sO:>?bd6.S֍zQ&4*`N|5Vg W a: ?I_߽:<})&w?o0`Q8h;K{6 ,ԊͩCKIvU߿Ƴǃa/tV"d` b=v&EaX ?߶#R[tMkѵB9z> r>rC?\+v< _ENFxiGA\SbjuV:#[ܛ$|N/Ih# ©rkq10|H  NA䃟'ҝцu)kOExH]npS'HcW_H0 c&*.վ kN8VnS'˦8L;N˅c|@C9h@ vkg hue.[;#sC/.Oٖ qƇ˄Z>6̴0)xMrDLJJ-Ѣ;imwHM/\+ypchƱѬᥖͱ!18fn|\oDZ~6vo3J{\w;;R#05*Cz Y@SJx->cm|7%Pm9b*7m:.W_ =_xDY_Lྴd~z?(? \[Aե͇sa3ɾf}݌fDo-6d X6=1\XyXFa2>\_Z\0-7R]KG>*?Hkxq)z9}y]][Zo6,5j^l;HHK## k1h؋9_`mI )ӆ/tJÃ&t=Е!)~GD~8jLQ KiR 4}J):L'SfE`ާLRbҥ|):W|&\.bڃˇo.RA{J %!&z!P:vCG{Dwczh⢦8(b\!p@%XxIC&(PX@ϒ!D!e!|Z30{-#cԀG*%"308-3ۼG .nOF_k)ѯJ[ML~H܈z1oms[4\77:2 0R<2*ܾpR 8k*x>RM۶7Z w1O2pw wϟ.gwoo?+}4,糵}ۆZ˗7n|hya4lioqu;Sy@X{ {nw |QۧB [N/nmK>,S߳} %Nyޠ R(Qh-77v-ćW>p[ݧSȗS_2J|q`*`gsw{jŕ܂6kQ4RN'!=j=WűOՆ;aR/1",DDꥦ0@ƀG!eLD:MȂ_>(܇Qi.ǮڡM#/tr>g?$?r!ݧ$Vt;( 0SIfQIp2Śkjw=4I5Z1@M AetۨQ1,%! j@C&9\8qrca'_sS\f,ɳG~.e5$F%yoKnӢ$JђdKk,)Qpe&]q:n>u'29!%#)\ 9!c) L0SP SpQ14&x ֑aSXA[R.k%'RHD# p]Bny*>x99lOewa }~}6]?<`wHvO3Gb"Q}ĔE ?#`SS9jߏ <&Re+C \ 1ay2% ֗HUHv $}o"d?*7K{?cx,pF`UkmZa&k ;:6xu$ Ջ> U>tǪJ,Q:;`C?5׶tM}oa$ /fIDt 0.g3]mF9-UpiT6|N-| 9=)Ma"/2Qa&>Q "]NA]yc\i.;Y;uK=SE`XO]RS~$vŸ?Ԩi`gR 3-|1rO+?>fX=%s0~ynf Dl!iorWGOo<}q4<36֏*[=f>sS V#,0м^#^CuvnSx!A߃A?qs^~jX72kDR跉Ÿ``]4V72\T `y0:X]Q=:ob  P_v:;6;ILh"Iʆr76l[ݝ-p*\]qA0UżdJuku^i%Q52 1Q Uګw_Y*g?-ݳ/J2Jޑv I`kF${iRzaZ%n|)Ŵ6gd;9=erE7;y:+a$D8J'mh>r3Vg6ͺEPzniI@ Efp{=K"0Y"=Y}'FϏi.?]m3zq-0O^'v ҃e @]鈥3(e1zixVd:ktoX67e`}hյPúqݒRdą{op+R#q=ԥx4TG2P@tԗ,w=R0p[zdL ^ QRkcGtO(Sw'eRw}gp:V Ҡ9,pC V*]G/ +`p" d1&աV@ALs$i 3ˌ3r&nlVpm%,KCcJ$hJXSڕ[iy)Ot-.ߥ-UҥM/bz(.j3#,q\#XI[NSbfjiH ڀ 7\%Eq47^hpg0"21DPZRzIPF >z61C r=fP4TJF&rȈ&*HΑA UHgz M3Y6qZYWߞ-b[&j">.Zj|`;a,[V iű0dtQ D0i9UssCN?g8UȽR:RY%Q[aŽIQyAlPMzIK->9αG/qАH #5ţ u82с'u#ثls[{[{TFm頴1["S\袵1Us1 H:kӶ|m: ee"|-yF`\0zBRYJPGAÀpV^S ݩG>X _ukV;TfP/1:ՙᜂ!oY @"NlHdN_J,s+9? * RJ/a50`yw*,8xm#GEۻdEf]0 xcKI_[%m+mI#ج*>Eq0ڈi.0[*bc#`[w4[f03SSS I3XfgIO[3bVzc\0Ug ś%[HրBwl+Gbnv'wF/yi}Jq:9R['{dVzxĻOvd RhMZhyYq}PLy0 }֒ Luw޾fњ6hI3XQgnD9!yF+kՃ7dYZF6mjBMP$S'x86YH }jjn󖾸0hbR^$uZI]Jx*y֤2D'RYۈ1MbQѠe\Ihȱ})y&1ۖW{Go/o5g#}刉b:se,8+*+VgˢwCbY"wrԜ1>B[ ɺAy$ Yq3) ZfI$%d f, @O6@%/br"ГhAGF53vR-q??(nЫɮ܇hZs-s19s% tr{i_jygKYWޢJv 4vϘ(SV*zo2Ogȳ(!hnP9҅;s I{Vf1$zRMRl.E3:"8%LNg&V iƮ Y•Dm˂{V4{fga@{.ht6=Ƅ,]ThfL9:@Ek9 1k-X:vU=6H{.  mAW(cfABU9̂6=v5q~2k!xjq(^ ^{@wi<]*鐳V""X<Ӣ%  3Ռ1#D.c.Y*7#EȄ 9R i51D&(1I{˫s?leYqǮVvGKeQ\RXMPlc2F%Jr)C漘; ,&.u)'mVF)JkH&b[5818[$.ҺΓ_-\g5-/JlU򋻗M܃(u|ŵL)|Q;pqr:#LtM(ۇ dGKSf^ie^fbݐh픏A;fNAE@ hA 8f@K@!D) &JCutvFnR3'#nU{0Gt(c{:UJG еقˇ#z9黷IxM2>Y0c#̓Oox>^ut[yÀW.5c+Znx2ƳZ@dKEĶ GWG]rɫ7toneWqJ'O?o&d~*YӴ Ia]OJc9E+b@< LdKQxC-2;zã {@Z'Hnťn"WW_:wUƽ7,u^tڻ_n_.N ,Q7o@/Y|_F8F.hڛLիi#t%jax"<-s[n3rW$0rlUW⮊h]) qWvˮ/(䓻/X/טuW_$di՗I)]/pWvpWv=}?#wE_g㮊F<wE_43kupWߎ$ȭ~mm}m9UPfQ'\];O׵tMy|)%07ߍywiyFG67\˲Ⴣ{v xFn\?7]ąg㦋.R*5oMK$|SAK ߓ4:rm?֑&Z#tO'gřFmf>;m%_+[v9?=khO>={"QGh2N4ԝ'it)kNhG4F͗=LcvgmBH{[8h>}[dO-40ӯԖԿr?~Bm3kZp-]P'LǸGsA&喪ҲqY'\Ɍq^ x"Cr>4UIi;ҟ3]n:ՓS欇d%f,[,)HڔUKZK(.Ut2 8d-*ɜ&N͂R$ƗLp$xߣ6s(?_RwN};TPOAqCl"ăFBs/1 9)B}N:)˭fKy@vv\XRN4W&j#K"Ajf}@ 2NBJ(0Ik@ $)L6)iOvVM-O_:9KQcVCMLdP΃@>eP,#K$yu5KBC0@MxQo᧟X["#^9C0hRc9cȨ,-""BRVU j' qxR0Sًg4ZNDad1ܖM2Er!88Y_^ O-ؑ*r?o3L'cUhb9dsDJ,\ ('jDYCVoFd>TЌo 'S2ބ(D́ G)Bq%K/D򬸉XJ v0zcUej9|2wOG4%͒+0= q-(˄T.R _lbIBGn>KΐiЗ  : aaHA䑊 Ts^d!={xLY̝931-FLQ`\*anS9#&.2!Gw^+Ĺſht 2G*@|J*CnrԜ1>Q$ A.eP޵q$2?{ H~ 09'vcqr )qM IVݯzI4Քx HgUu=RiI N=Ȑ$Y߱/tDFtTHc˚߂BS&Q 5`DpJK;78|vr{T^dORQZ}g񠋕d`էy#MΚN?O|p\SƎ 7xWnǃ w0"&jDw:H84Y꽢:0 /i#JXj YO̰IcƠSb-lB&U3vX(c_^{^n22Xm_61S?w[?mtRh퓰Adu{H \c(  BRL+{"+ʱ pg#9Yf\*/IS(mND҅9vln< wDcŨXv`7Q<]2*k %X8X@I{xDьEa@Q~T;-(ȐJўG&D%L{y>.$G:*[:VTl燍PʌX8b1ؗ#爛(^dE)F!#)s d )>Q;$D6j&1"lpjT@LN!(Q{4ŭxaX;;8"^F\j|qZ71:Q^|)|]Q(ykcA` TYCyF#ʂL(B21nzxx4KQǾt?~xHp;ū)_qdR%2ikFx `׻<{@m<..SD*`r$]mp6:iHYk{{C;Gǘx/{;#w$Ga( (( h4c`G.4!& <bRetUՎFx:s, `eN4.՝b|"]]/0۝5¼ߏ1W0]g«1IO%;;Xg]rs*֌$)ԯ-g{A`})SxERƒ-l2(d.TI'F/rJTrM;OjmHx~nz~ ^u96Am:^OOSxI%tu[zs8Je8MXƒݫ4ռ*yjSr3=L -n%Wcޅw2d=o8x5= {(GwF!ے4_5\GyZUts,WL$#8-/[|λV**gȴ:>[Uj-M90Pd8[N ʢ47sVե\rJ@HbUVf0 xW=$k (daPN59PsNޞLgNw$1e>!9P(vqڦnnq.3,wclC hN6L),ŧy} :;[C8b"_^Ja?rh 9|7]uoLU[<c}yr]`Yg3"+i0Je;ܬG?̚l;@HWKwk骫Ռ$:K,oqӆ87 g0b>ǣb6WYU*#[] ծR.Y:͗qGrULc5NUfV_~^) J_uחHBvtݫo_ޞS_׫󷯿O8&uIW< ?ށ5+Vs~Ɍ]ͧoF{8!͝@X'.KUX|pfL+?sZ# ۫v_%K5-vhZ.B!wׇ _h._L|4P Axh o$! v|tM|Yj\z{o I*0c%i0VЬcCD02pC@G!`3چw1i iv(k9u. ~B {@`j)r3룉Y&K Wh[, :2U/q]/d+?E?C\q,z‹yR.|N$>!^'0;a67c8s!*x\w5}o[FL R.p_4=cr+jϚC^wk/1tyq"&4F: #rAm ݺ:/zl{,nyV0vqUoGͷ_6GwV~6crbo6%~gv3M[{ T|_7t ^=V.'RcuزדAV|Xz$9J-\+n_)dzSȁ,cZ I訒V5XICB{dNlШԣkCͺD#kNgcN3e hq146AqUQ@{_fǹǼRymԁˈjHmxހL[jF/0#pFOZ{z<. +qϝe9kdmlH5#R>ibc}RUpAC^ҷkhqahi3}i _,Je^6Bu8N.^]=ꚃTy*gR"60R<Ɗž/lr=,Fq"P%Z3+a`E(@4_~W|,|΀oQcgZR- gئI Ñ,?YFLk^%vֱ$'L8=y &XJQcQ49. 2JI1Q 7(BPy kl $<\Iʔ㶀xZĜSÔT'VأXw[e>Bǔ0gO!mx=Eʼ!SB)U0\1̴ Kb3Ac&Yl<1>D0:1Ee& h=9PSyOѹ#r GTs 2ahe !q>2KdV[0GUˠ@ޕqk2?E;wrm{oqMr hT3E6%gĶyˏ%?O$i [&orkM;nm|Y|+]E&a9Jq/&Cq9)ԾmڞI ffיF{&wǠ{pp`UhcL3)vA-bxs6/:!Xd<0(VyeDL `ZQC"u@Eug}j't(m_/Ǩ4W buQo?䤌]pXzs_YuWt,Lt% \8N*O OK"1rpm#Ž`1!H;jeAeC bl iϐ8,<+KekT b 1#u-W,Pac҃ȅ]tb*)NJJ8lNJ]R4v댜alj<5(5pf[ktfn_Mmk^j~s~8W({0Vq '4DRe b^Ce!)c˜xj:;v=yP"!k0;`a7!`F9 '3B둧캶.Xt 8=2rgL/уE JFpRsa?^cj ݮxw5:|ӧzXgot8Df:nu-ٕzb[ VueO+`PK .51BbKkRV!`{=%cgt92!Ω"gm{ BؗxVً4.ͿQ;.G}z-;V[rEћp Lq^Ȏ_!ePM2p0ڣdxvFÙ칶>JP;n4 f>[AwY9R&qk(Uf?8rTZ=^% ^VY?Cqf\M.Y e0{5&hdML2\LxJҙ!3D씫 t z4/Lc7i:k#pt:Ak:_*?ꑎ ?wzlL܆Y'arѳt98o-Ez>~⑰pe/j|לh3VM-ӖV@㚜r 30sdE7 4ς`/M5„L8潋D'$5Ҵ Oլ(ogvkXBҏAKP#}>h [\>`&j6 Y3o#h{W!Ro I~pdqyٮr󛛘K~Ңj_j#P{[ Mb\;GT;N>bFtOHAZwCq}MlM `2-/C%龭apM9^Y1@mi2k:fm1 uLTHZBqvVVDtGDw#] _]Pq ("-G60 !$S!^Ӊھ&$ӳO蒮Cx6=Y $")8AR: _NZ!#ߗ?Jlڅ콨l "Q&b/i#9.3ޠqEu܅eDF_Pxʮڡm>ʫ6'a~xp@%lXn y}\5ڒQ9^tEi6I|:@~Mżbؔxz1~1TͧN{"}bPa$7J7Ї@h 7>Jl:tM'<^n1vhozM]sn2uzAj_,;ڿ˷ND|<|uzMգhD Y9Jf(IYTOB7&j#c2򲩠CVq;)Ҹ-ɱ0;c/}H2fqSظʒfxz~ `B qtaGQ.mvZV9VhC!O  AA{8P0lGXGAղ?ۼJyp K<3};d 1ZE kya׬mKݻ;xA f#Ln6ծf#mn6RAu_nV2(!K3pUbW ]Ppļc+ J3p-PˠUB*L\E1rɦ;ҭ"F7z+ 身~Cȵpu;r@ Wr(pu;* p᪩]sJhCp U$W]+"pĬEq ^_\fNڥ' {t7,n_!}_i3&;8Wt@})ٷޜ $6l0Ӏ9_Z<ʛ+n "?r'{/4@0gL7lnOc|[He-bfxۧyM)q1 9 7E'2bW{Xa2`Lp@.A;#ˈTJs_!A\E 3pASw"n;\E*iWW *.qWq #~s-^V'Ӄ==)W]b&G0 %iq>ewxR)Z17y; ,X˽LЩd1{.I$ùvϽy/A0wMFWE~1 tvӜ<}0H.gڼyބ@7vL~0ioGl#w$Ea|:,ÄX%00FBa$Gvr ૶ޡAE`Xc1[  @e:Yg\Cmt`cqM:gCo;aY'+s&ȟ߼ IUp,bR)Y9IuG0.0l4Nrl!r͙F&7ڐiEzb9o̙xŎͺ,HFk⅔I H` e!+dXk[ f2||U7|Ez.Wocio}`:V:\}#>e>ja{Q }>jaZG-Q OPh)WjlC[{W(#li*f'®$x˛oVP7osv5Ү?IoպΞ%QFu#.Qkko6I~"C;d |g #JA2P[ ɑJ$ɊaJ{K/a ٸ9(#I btݐyrJ敷Vd^F;ݱ1Κ@DUhcLysELcS)[ӡMjNٝqwkN*O OK"1rpm#Ž`G B\2Gd164g}g1aULcJbSPL!fd4S` #s=03=l<:]Z3Vl*WPyL>08 j)<)LU )T!7[B;m czxгzzyjCB < Yp![ k$ 5Id8!\< Od׵uJCU쑑cwE8\`D,VQ*0 qScvb 8uf=Zس7?`"3tQbJs=P1-_n+]n '~k0pA(D%XrP 15P)b 0=Ӎ+z7Oc g@sq6Q(e}2!Ω"gm{ BؗxVً4.ͿQ;.G}z-;V[rEћd\&/Ab V@.܊gh4):̞G (~ƿɀ5Y9;n%kc<O5 Yn"X#4_|dB)em5*q9*eg/鬵Er8*W@;%14rNW*[d4ӳL;M OI:3}rU4MGC;)"S{:;7  ZӨi]Ri7QTؘKW>UOUWӣg'0W grp>r]|KW++_?٘ N9]sX5IL[Z/Jv)E ͳD1lˣcSM0!jq56tр;?)bjiHLYeQ 64 5}99GhJ: >F1@P|Lm@`%_Muc.giڂ%lV]v!u6i|ͽbLeH+\M|Kr(z|cˎdg_[lLۚ&~*u?Iš.OU"05y=ݱ ӁMܴE5O;fo֩VDD>v˔(l.+LFT Hdo@VԺ@T`i⶟nS|kӋOjI 3篜p<>N<вƣ} R@[f"_&DbN$ATKZc^T)]\?VXo7bMԪ;PlR^߁ ­Ǫm?i'jNM#ho茋jstQ$jk :aH.tАrIE񥊒K%+J^DB ! y]x[ ^!{hb@pP9) H[(\9A*f* ]3rֈ|]*zb:M fU/7k@,;xtM&dGA{{wsx{=w9*jW5Uv}+Ȭ,",0#lؖu5VJwkUobՅ<ᤏX`ڝIֱKEbaЩ| @ةvc%v8jzmF~ע9GkuGsщT/EY{m ELrɑ*Ux,Q *'\tX۾b*"godES6*l!lFΚum W8D9#]l&(5TZ&h̥d50'c B"K)^ Q"K@46~T Xv¥AJy\Ph3^_>&NI3_mK8z3>46G2^LώWA,uyNfSVx-ͤԇO笽bbzST^p1%^YwQ.+k2Q,& Se$&$ْsQT7:tICUm26U]),e?4 TkdlFNWi8cS,t;,pAQ}˪Nnyշ^M,.'hKń A'E!HgsPB>Y vM[;]{!+cTZUhBDVyVضDǒd]cnFynۂn; !g ɦR,`A*a FYK`N)2@ۚŠ ):#-d %4YƚL""Bdب a3rިǜU`Dl"o~;D|Me\R*e :9W ѠƤ"ewD&e2۶N 6&Z5 T9PJɥZHJH51"6#gZ4DP6:qFz72jCmM6x!&%mAe,)ɪL SV#qǦx( ~:TtG6G\h E?Z#Yv|jj?|߾9?tZXձ"=uu<YcDV<ιc^1&R=%7]c49L3MΜrכ[78g;3aL ѷt_<DFwWo4\v8:Gw^d`o1)1w8ScN'aռ& %3[gw^7á_m\0ѰUQr2!?`7ߍbxe!6Uz`8:,H%HfU}yi2k}$޻dL/k|ީЍUhoog˕r7@S6-&,BBDt9\Κn{ZrO̶x]􆴬+i5gOؚc5yj|Rn昊d㕳*9&D2mq??@RXݱDtH.z6KiR fQ}1tڷhrqaW$HVBSkUUVeҪ.,Oy¹?֌*ϬV!7FD*dH&%]|Lh6 IB5`{e"H2eIn3ɝ+k[$ HNjx,:u1 1*Eh%lÛթ v߮Vp%hB/$32 U@yZ/4a*\qls8HZY3r֔^l?F/!Kk-5 !j<y81yixm&jAJՂz'-EOg@N2dXφ-#_j)9圂 wB:dKDX+ LJ';;ig'-wk;:oI \YDaZ^uPY9SrAjJbK.G k>u} 2M·sT%K7:,KP;JIꔑ)w#´]N;xp`36[Co7a Jz"Q>X׆!BJ $41cL}2ID`y@GS邋ZLx z˒ȭ @5#g~= >&Wl7ώVLˍ;K^|?l_6)@ΗH'IAJL>z3ۀA{k4ԚE*J}"&g|K'56 aglm9)5$NYHX'!@qLβ1$FG`=}< UgPryeֵS .Z+0Fgj hT,Jm5I2 ͗J Oi/4#HEbV9,|rQbs)\]@H%V {Sh[G F%)dQ𡘌p?ކq|G?^5vCp(GI@Z?F 46ؗ J ?G:NtN|7;7PP]RaMN$@IXG4a rq2>rwXgٻ6dWm`TQEl#&WkIʲ俟UHIc5驮cZ l%Dk{5k_%NNifJ|bEqq<VpYwM*цlFMrߩōg.&ͳLuLTx}5^8]V[\/^v/&sbnDE{s5 ՟"'{4 .auHBvefy6^O8y.u&zs?kGedY)G@qF箯F"czD$w_v jMyNorPpO?o_w?eӳ?o8HQ L/bbbUT0¸Z?o;#|9*@vE,"@Һ 3ÝQ]Nyܙ(8lha5il͸#׌q/$ 1*TB%`CP#ځo:8u*1x&qJ*lri^".&CSe {60טV*.u T*a$6o@ -j5XD+A@ Ave"^t+p15\j8{)c{8H=c#q%ք%; $Rr] 2-1ƦnomuέMb?8ݶ,}q4}m)gv q DZs_.hCL$]TwItOZ#U!2n ]5ˑr8~.bsJƻX\"1[5!VȣmJ woou~n>׽g}Uz0#kv7˺g;r6[Ov֓\(tsZR&FZN /n'MK f<᠛pG2J$F (̊hX*]Cj )ahY{ЪU/BzP УF,\wU*վT* 1~Pb)ezLRӶx)!#YWE/M`;u4ŒZ+ƍ[Œ1bٔLjnŁC<(xX5]9C"e`T<j[D2j81rkhk"%zuV)c1rlDGכpԜ:21`3#6.}65ޜޖȵcYʁ͸m}׹׏ztd2$]}%?.'wmtfs}Yu-KlYNunM7w>+[r=? 5Z<69ktbfӬQ.DŽ^4k{qrbCT:d2rkM{2%*l[Tv#o]J8u&{ȧI(eBekͦb;%phWсYF&rHh_bw %p) bTmAL&ʌdXD#ՁHO{cl݇%`9w@q> ;D}aKl6jn>j|1i?f{m>ϛl72.dErJ8CM>"(Q`p%X*A S#  %̓ KmHA4[mMr=)ZcKnqQ%OG9WJ= Sa'Z9(Y+Zdd`vÎʆ$՜v$(+A-m0IIV<(y{[* eq;C5B3nR ^2Mn=+ѵ&8p 1U$ aR s(#knuep1YhWK'AR!|Nm!̞ ㌲!Lso}.')]HV _cxB"K^K+#B EӂG5b=vOZwE8#rTPwJ;B$JID͇g{AN"[F!5a{6y*e_H?[ݎ[ǏY߆I|HOG UAȜ@+)\V8Fyf]F9#6pZy  `x<@<$ Dh9 sP%t`4*-)hB)nXz.V3~&/8?ݭmb3oNPOQ>>.Y!4IȭGG4k TJT;N=ʥsiFЊZJ&yF(̥?=mKZ"DyÁo-A((c ó1'2@7Э#8,9P"g$yBK`(x9-indDdWjQf?P2oRFvYlH_jBjJARMC:m%ii ge5ܜܩ (P6Qd6\aN" +h{滹BkL7z$!5y8%6G'44]ŗ?m8vm.^ҿ>IY'NGTs Y|9GG|'o'2Xѿ!O'FeMW񇿊%̾zV'u+'{Y[1vF(s:*ohQ ɉ(zs~3|!~/ ͽ~/./PAɤ_$}0#~G(`}&,ubw2im3A(%G#8DboB|#6JT%{7WKՈ_/0v8իl"SiK0 "l8Ìߗ\Dߖ )Wd j%>/'˱6S[ggLc]@9 }F,Æpz-} [cO YuTLGm:_%WyV?ۄK6jhG?t@Cnȓa{]P^:;?ČP1 S!g0fSIg 6`&hp$M2LV޷(0 ۄǼ7&l2M|~2of:KfLDvNt_; :f{}" 6.ǓP% O`itI#*&)  Z{O#f!'M_|RCN ^N3> /H r؁`D ! KqAHYV !h+z}:/Gq|5xjXo7nɖGT%7<,zz˪'(yY==sي*PLz:3<ݙZvRvKݝU<|2/ vB)2|Dio4VEmK'5+"}+ANy}'k>@"N$(0k (O!%? R8+5DC":k֔z.hCL$]D(3&H|200Dfv >gΊCMeX+ˬ#歐_|͑I5p}-熮[0'0DD;RW^-?PbQHT̖S4"QW9*J7Yld9Ҏ{j*R(UR$FCZK}#zzJQηu.aEGO>!/@ p1(0:yK4NH9d:/t\KTI{;ޞ >MjSMfsCrV-Awg1w TrK-wmq 򒓇x) g=@%E d-V#VSь.VKeVKiɺ|UK`B? S# ,BL ]QF *9C-Ur?'⩚ͯ~tW(Ĥ/|Qk\ Uj5>תYzO.ƒӌNvqsܮgg'LJCnYÜ|lA>/i&ӕ/MU)"f좓7 eOHdYނIʖ CSh+%%Q{T,z&㰄0R :Bf?4 ֒tJ3Y،2v~/JnfT"މf32_޽8.~s1-r@بCB[ A(Ky!!2k*xQ ҪMm],Sv‡xlTDvn6%v:^5s,RuEjIjO>g˒˥8S؂M" (1:rBeXBxm c] l!32VtX l)-GQALX F5fyXF:b ƱHfD %b$$xS Te tZX#E" $2<_ ^ amL :F Tr&ab$KV:M>bs NBy%blv]KKrq9&_g3*I.nJ5kYLڇDWSǤAXkcJۂ8R 3-JI.>\$I&&,=[dR` !־UD6%.tlRc'yvtl&8~9W)>ښ|k?G@1Ww?z `VHĀW Z׾l3=VnUUidR"VQNIZa*ڍ0.v3yێs%ZݍLe GJ|4RޠBWOk-FhQdJz$e$ؖd=if"R%!Ӓ̭Mg|#:wbDs:a'_*/i*!z@;J38Mk_^_70EdC2+]-kLF !YBIl&舉1eD2e8vUrG pg5 ADx >Y/`BA)t* bAhSk`܏-"JNieL$=dM'T'6gU OTk&"ݘϚMgK>?6[P७T DmL̻,k9êqR.1"˼Ղ-Y߿[g,ѰeJ!;W+H ZQ'H!7hN1G +W& jd'Mv҆y]cLPB suT)QXZnFĒْ9!Ip4joڌ&vLЍKhx0FM1jL6ƊKJhbFԂ86՞r ]22'Sm<"moU zSH )d*RB4*DAhl>hwʘ 9p2@m 6>9R58^8ɇ̳' Y$ec# `:t1Wta`[nvt=gua˙DH:,LRR 3j)%udHƒ_ue}K_%_iY9K9K;3P|wR:bdY608侖{0ڭ%7NYo0Vu抃ZVeRzQX&o"蒌JBR;Mj[%8Qqi"g!"OT6TE1kG @XUAe N$dz#iaD#_7k㈄+?TH ~۾[u^ ys5~Qo|q͗{yxx5{ bͫo'jW0ٍw{🽫+???;u`dX2{EG?V$-? 32mwN@ # B0׉3-⸗Ou?^/ lxT?8~ɡQUj:/G]LZH04G/iv\fUl=9;^_>tսqࡖi˳|xrGGx?ûRow?~珼gkGڡ)xXXss7¸_Ztpk}Y~YONr3M3\p:18\#%mAʍ$*Dk#pEkʟ6pl'q d$}koZ/Kp#F`*I>cQ1%4x uBĒ3 `>Au*]MNv9y񡕷^Yށp i◮,>oD؁KQ4ũ$3E4G4U # `ARIp|blѕ!IA^P0(]HdIɢB,ff9DddF) _ #3/gQM$T1k6-^W'|1 ) e21h-#4Qb)""Z$1jW6noC 8H޾ SXw@T!*MȊɕ@!h%c|f<`$'=>vUy#lZ|J^D(caE]iGܙ8qXѺ]JTnc}(Yc, ^ љFc:霐3ګ鿕 ܁)%3r7/a^?[c3}3֩lWwJ fmjVuskm۬WIHo풮OqU{=. aOa (>@BFYt")£ x$TL+&_|_@}Qjwסo{{s>U|>>c/j24t@WU'HylvۼȽVC[ڛ_ǟuwOv_f?Znˡw{wt۶=Tm>?m܀t6u mnݡhW}턴 iii@=Һ4L-RӓyiqֺK]Q: t{ LEzzQگG-~05J'HL;|yl`)yOup&zGRP&4xf܈(کc]jD(ҊL&`,bUa,X^z m#Tk+/|2NmHe/#_H6}lU0$%߯zfHICǀ%lTW?U]]Vc&ye4zl5x"uә8-D+|L[sTuٔ[77fcYOms[t,9y6yvlg2znu1}6:m*,ؗ$ G5;1࿋ m>ɐCZZUpg;:]ͯu4e-Z[wmx|xwu=ܮӢԃ{n8h~[}tGބ˪Xwサ-YK\ilB|vMIVz_i ߑxR[䴯znoD77ymyޯ-2҅ӱVs%>WظAXa:d͜8| مKu-֒6/=}eJΡOq=^9v '#3dKRg!0 'P{Iz t#ǜ1𴱷=Fd^S0x0iklj֒HO#" bT!AP8e. B2/r?(68,ˬ7ԨZ6~ٿYT;.#8qJ(cpC"h`.`.*-rSOued =!"}ЇDDpC`=N#h6kG3qvO%)F{jekf`v`8K髶[z|WA8Z>BwߥR\17I]t#{?+ 9Gܷlhd"M91/Gcw6R%pP#BXk uڃCL޹/߇_^.JǴc4q|chryN>'!TXO,H"1hCn \x눱LaFDI2:[X˷7 3 alR)4aU`qJ)(2pG:l3[#w-穻ͽU1|wfRQ [sP$:j4X0Nq(TY[1JɽtcxYNaYB@6H2pKd9! 6d|({[t^ZP8siT6[[$9b2I`,Y0<3ާKn`!x{rRљ¶雩"tUSu+da^Ð/A˟ Ͼ7#3v fkJq;*~6'礘~9x[lB r^~'x!_YZ=klx? ZsRX1)ooUR"]مN ', 0<^g0< Ȥ{y\SO68Ti1/\1J6ޞuI8 jpz8Y ;h [[Q|Ob Pr\j^=ǻ*ݯ]tqo6VHBtGD}EGEPzniI@ Efp{=K"0Y"=gǔ6:H4 VÛGŵ syEm[k-ݧ*J ۮl|9ZazY9tvHӲEݧMEFCE87#$Sq#Xm4n?2mS;W-N۰$:^ت|7KՊ1٨D>Ʈ1VhMpkRij1/@UAVشۛݹW6~UM/Pr5VYnGeLʇXuY"e&`,Xyv6O6nu<| (m/- _K`I_O\Ο~Үp ^R*߃~J`,Oq+\r籃aTC]*몬keI} Mz6ŧqCxH]߹ysM&U [=&jn^P<1|ϓL/3ח*^W*kPΡ"I_A 3*lW*0'K7Im<.bviVDtQzh52"\0W?HNZ,_HDxC $ryR"NRJ5OK+>f̓r,3=yhh]6^ tr^fḱ˝ʞ#۳j7O?{q_!R|@V33:iSCafMHVwfyĕȢ &Μ99_J箪\GH"'K'r6*+Eb_<~X ;Pa X]8EZ7E\*06GLя9OT6'ˇ*S^\Ӷ\<^Kc3kFn).4MT)|hwi|}&[r=zwS¤q֫;N/tJs?NRMOi> QJK- Y\^L΁y؁uL;eNC$swq')eU\GVOuV'V++>П2JK]-7Mp'ro=V,T >]ß.|G4X8XfY֗~Oh%b_]^Ub*FU7was¯CǨnJ`"5M>ChP6 ;]D = wݮKpLZA)MY囒Z'1x0־b")eF 朑D§]p_s>]3-V=8=\J|{~P7}ZZ1OСPц+0 ښZ`ZmyWes᭰O֮YY{0x=D~\Ѩ=?ypǽlxR#mF >|:P\ė& Yem@E&7,CI{pCq(.r3Xi#*.ŽHE`K9@,L$$R4$SԍVˎX^zH#``:BEg}N'ȵjӣʒ/S܌}|8 Ҳ"ݼ+}pw;˩jr̡bP ̲nޮ3=SpZvԋiYRZvh:ʝ@eӲ3fT7i9)6Kl}=ro"sEql׺ C=,9nkS6o)x^k@X]LS `/U2˒ȹKR.ѠK\DMѩHXPuX4#%ukL e\ʠhJ[+ [3Z|>S΁( [rU#`ɪP@g@!s$F+<\H,dYڀh,uYKLa2nw',;a2 DNr6g-deB"(0cJI0gEe:NO:k_:{w/AKDMLҩJ:eIG, N9)qd5K>=b $߯W~f@l=*l.W^ƄKT"1,0"N/ʠjUSb2I!nV)dpEFbrcH: #GctLKrS@\>F60*cWGSNX*QTn U5>b:uNDC4!{HZ%"KBR])UlW)=頏~596A!B: INh990Τ Y(.diCd `hN؁AGZ 8I)mYq,$`y\ֈd!9RPCYv̺ӭnM|110Kv1+MLǕ߻KnU.|(| FIѿl,C NiO vKr%]2׿wH'!V~tgO41mA Fj)$O(Ɓ`䂝3ZOO b'ibճQTDYȤɁ %R=L`^p #ln}mrۂAi,b f6s1őzN?хkHF{l߯Z,^z͖^Њ#b2o=MҔj~ë]0ޢŹ4 h:rU+BN{XL(:gjL]i4cs]dyMDtٲ|pz{q4+v5kZA*\K{ZHXyvW4Qap2T{Մ-OGa|Ďǟ*?oWz?#bY % ?gcՖ*Z3g6̸Z!ތ~48z})‹ ~O%FĭGq OpLGa5DGl˝ ˫_ѵS =FVs+43y|{^r+n+7b#A{I ]Wm ï/﵏nje#QQ߸d"N.KUNAq1Ȁ1%fJp+# ='DǐfK;M6 ^UhNCJa ďU>G h]tAG Gn~b>EIL0 0[U%<㤱6ShCZ Fqt\CXd9qMWK%SF 4vD4V A{ZKrzѶ6Π LnPAضb]ƬFDD_KGD18-Hg5DwV8HuV9»8=䖼>0 nn ڎhwmn%Ƴzy7"|ΛAy]^IC.혝5@_ m]kƕ+`MS]#k{XUN$K﹗fX#"qF,]zSϹ%¡o31?@KhxĐA RYzx(@mW-2 eV+o^#"7Ro|OQo|L}"[]}@v삚f\Yxmo&fvP_v;(wk偵5;cӹ CJ{s/AܵݭqqOO]x'C&8 gxcVV}XpkqRЛ ^pm`o#^XETScޝ ,D,85yA[.*(9"`<1v;sw_A›b~Ay")N..MVrZW3yف Z:ʘ1<YrΕ'|Y9OҎ9^Wy-\Ժ@4-r.Edó\tQh eT.DA4tU'> IV=|Yڛl-sqn=kWPޗ /"L<*_91$u"S,J DmO@X)&eBtm:\Nt*th;]!J>9ҕ62T 2wB]%])+ Vɨ+D# %d3+mnB$CW%\ ]!ZNWҰΐF [A++KB|80+vdճ1T<{rZt.5vh? ]CP+ւ@WOzmUWKh33lhj\bj V F77hGQ \szAv/%45Q$@'p|M|&QNQ%hkfuwU&yѯG2ͣxSw@ Yx@ mS3\@|T7Wl-521_-LǛzx94B,8k䪱hS,ƕ$$~j˿p3P[˟B!fV?OBFguzZ?fG2Q I5JF!\fSjV^!JMvRQ¬H0W<T JwBJ tutnEKʪ֞ Qx!m6Son!cVr& fH 9,Mhu3Ht[N7ޏRϔt%|LygTbD |F7ֲ,0,RQd2Y{rو9}:!ˡ1EB"{&}|#{UJ2B2h_z QځΑ${˝B$CWWT њӕd3ҕb:&}+@)0e{t9N%DWXCWWT Zw:C25K!$B V#]YnNifka+Y*thOuF+>GV=A>jXSvpى2C+O4oAW|V=Bk]}+kLKy>]!Z·ph1uݻLߡz)\[Mwn7hoaKIU"W]4µɨJ@;M#Ja>G +- Sv6_vÊ] 7ˌfR*^dhoK.5]l9 n7\ m qL82B1b+g`˞Uq6P/R:䶒mВ< Y&8 RddL8tZ\HBOE"SA$9JB!n@|.]`Jl2tpӡ+D+H Q6|JJmlJtЕrvh9;]!ʁΒq4.$3pO}y;N]!]iũ2 fg[}kOWojס+FEc+J/&&\ HQ--il}Nո\n#7rIxQJɨ{~\ĉ'$04) p_0RQ*2H3 Nȹ#S_ +D@WgHW>L8Kպt( JT 3UL?8~v(`ϑ4cJµɌ]ZM{?(0{te:+lMP=Վv(tu>teQZ'DWXt Uɨ+D5>RGVN9Vp59P{;/{~]Cٷ+Ղ@WOzʥIխKɓ+e*th}+DI@WgHW1hJ&Rpnv@hmn, 44FҌ[BdYQR=94Hb_.cMF'Ḱj\7IVl7KZЎ&z5*+kbumpCGrX'(Ǣbir7sy紃z_13.˄t&s^nj1WHQ]EA|؊*?T?؃W77[{{ƹ>_, :˧3 hvG]Wm?_mjwpY PP-||b]ƄRYAwq貅[ƪ&wƨj 0/=1d:e1uVs:tfwnyq5fRf2[aȷH=,=6Ow[}̓?t:VnvGr5|}=]8k>47$<dwo=؁VуZ]+/}to^*;Nn1mev/)_j4Բ1UfLS*U3]N{qY*=w". {F㏟l~~ڠͼtnnl *<{(e Q缻B[/w2ő <1xxo!bM{(t^/V )c!KJVh}jF93}NYp2xl0 ((b)F'hոx3o%2xt5A C&8 gJխyw{;%:p0i9fks\r!V@#j$UTةwg<(%i!  NM+xЖ ʬbN#5(Os{ݗ90b9x9wPX,j]f6dǩJ讌!s*MCǿ~XJ~qۼ`Een FFcZsJpFJG,z]"NBJ!B=[ HLRe.r@8[Q^qu܃;s?>~ o%ۿ9>)*+W 4Β;Һr}<f:H%yT^&f"7dncw}K,s$|/負%'bm36cI"1Ǩ:ɬryc"AM ֩l$*\V )c$pύ][oH+|],h"`2$<B_mndJ#JdO5Iɲ-ZMr-]lVUU]]Qʃp-a`5g9{~kmz@x+6b%lS3$6vq&JzlzvZh謻T#D("HlGS  ՒFYLy H!sr7YV5k,byHFǤ"(fTi%1 @UAn LxUk= eh]̧ WuQ2ƍp#ԨU0-'6m,(BS$JKϥY?TK>v* AWgV9Qi i(y-GJugAz_Ywapg<iDA/z;.GI\92h2Qy}1eQK+_7Nk܇d'^fƅ"LLol&=?ߪ;] jQ ?NÌwf&?V:)m} .oy"v5KB*|t&a@r љbLU ;|T-@^{J2a !Vr m0oNljBa`qSo%q!|fV"u}Wkxq1;޼:`HӦ͑۫ ƣqv 5I#GߗNK0y]yu4O5X~nTx\N6 b@3Hkb.E'y3T_s$+eR zr'!5R:Gt #bl2[Zi`,U|ZN>&(Q }Cv&p=ja%#i`t^KZV!,c|ifNlng; ~>|j˿4~1&j|O`"8XH$?[&_JEmn멥4{U+*7˟OIh}OV*R_}XwmӢ0-7&Ǔ7)[ Mch| p}\.|B9MZtz>+[iTkec֝ڴ/n;Hq{Wn#IOaS{m8N.I0 $Iz$5=HOlX)a \ ;h`+&ɳsp{sq1ϰcetzөb9{ ' }?˰?}ma^Hο| sSX5z ZlCR2e[xm1\iM a* PcQ<E C`䐤 ir4"y"_J!LE M9AʼM'< Scˍ;6.LrJ͙"ǜ M2Ymٚۃ;}`zZFS8ϖe--;_2[ի~W}fkuT}QP0vY4a#A29HysQ b&]{\TnXo3]F7]N|ؐ0P<% uQ]6Lr-W߼rHo X-7`$'S,]v7V-7od27ͥ kyoᖅ 0|&PhJ3 ֓  Ƙcb>QC{A;<@rFϵ,wJÃ&t}ZEY=<XV>\u㨊O:R=R\3Xun`O <=F0y3ݓg(eגB?.{/r^>Z^3߂СRGѮfLt޽C28ks{E]*e H)n`+,.ly|%?_`$ 5"蝍T +5( g~(1aWy^;oE^ǤqAX]=oCQߌY޷6})WϞRb9>UT qc%8r#zF0`)q!xE#$^$4(A1 `*/%% K^@w9JaظpG̘v5|x{op7*+@|nwSwv[<<>>T_+) )LSAJ$UV# *X1J,hkӏ~P=g<_ L!DJSCb>xA԰A"7! F^" ӋIpݸIktCq?xRo(oE843`M62 D(^.,2Leb0t{ߠޅ֫ТCf}vu'PԷeWB4w 'ԇO6ȫ_#aCBB8b]p5،Iu!tZ*eDa#fPwW){fbln1!ɩ0"g0ln, B_rQLFwybߖQW!;^ȎA$;nFuzxz#&X3X i.%4hNC?Xt'q(*{f kBb# &gb;L"Ri{™IEiGڋ2siV[U9=.YkodMm,2}J&2 孃r͠Փid vo))%_W2):FmhԻ/_Ok&&%~~]ٯ&m\j7fV_#x9Gx6oL?q)Rԇh_YZV_w6;O`{>]=lqۤsvn9q:g\FX#ayУP| /d6-=\N˔!CUYI/>OS7߆ׅ K$c&ͤշbvPGl::oo(YY.>k:/aF~J2Wwg|\;,Lδ Xڑ~[Von]G\ތw`}ptFTMuz:} ;rĀ $휌sM,G1rG$[Cw3 ZwޡYYj{Mvúo>2I8(W:&mevtm~ֆaU]$C~b>2qXwqgשVHDGD3Z9cqv|46vsE&[e@/qĂvl^sDLidO/nu//>M 'X7E3u R2p`AY0tBh0#F:bL9Z^/J/qgy_~Wzoqűt_=Zu4Ŭa{ą4 t0?i8\粹bI5/Hj"R$5Q+ġIMT*>I 2?Jo<-IQ PdL,QBZX:KpM'-WG|G{ޏyM~-nI߻vy諯/m91d}-d3bTUEO5Ǩ^[lN;XH$<{Qw<M̽5XmU@g2@E-8v`}0բ"hXPb?14gp%!gӼƵ Zb'ֳlg^U=_@6jJu6*x8) z]mLJY1 L($ xa3K0VId_LF5P]ML|y4ItM5u $𠦡!FS VJP!\Ωcة(db('SD[z 6]Yb%-ZUW@Vo!`je]-6ͺaܻ,*hm{ uq pi^G\J,BTӶjmϦ+z1pMW /\;Z ޫOtH~1IJx4lZpvv'DIb֦!7~`KnDjvj+ݩںO#3 AF,fKm7FZgJb$3gF39Tk=0pRv$6StiPPC1F)$4&a{){ S&htE mEcmUbm6$yم"f@Ib*04zgcUj4suI)_p:r*8|<oJLN4qJm".Ͽe廿Mc7l:~رJ}Y=)+?Qj֍XmzqP|POܰZ2 HÜ ZV_} S߸zgQZ @șyw(`WC HrIG6q'flgR'ŋa\s2)ي)yf^yq]YLE8`+B1eU}ATS\L2sJ:Cq;>| [Se=ȍ 1^^[Q1گH`#V ޏk{񎩗j! Q7?y2A$6ӥ`{n;ԥvK}CuE(k'Ď16:5lxfg zi ~9a퍷}\%4L|'zZmSS٠4݂wÖ$/Ϯlܪs4ufI_̽^~<\5bf{u(PJ]O.p %O.(eN.|Ʌ-#R¼7tE¾UGv:Jv3]AbEwҕf:ܰ7:Z]+E)3]A/ nc]ud:ZyR2CW;Fп:]=0+cZW~v8V^2n=hm'0Cp ]um>:J ]c[W+Ǚlև<_l畓t1ËI}o?=ťEgHg?.(P>x]թ=hTU+{CW{=HWŰGtu/tѺ?(W"yfz;tI~o{CWu(t*{DW xWSѕEh;`tQ՛+ Oډi5Bjå:Q;OWuE'RǺpv~OQ[+kXyXW۰tQDW;~ؿv vǡW dxJ;1ŏ+jҳ7g {qc5.e ܟ+mR+w ޻"ufWIgmEKnr W }1gtlh<=_-IJW#^СcU~Ẏ9opcF,[o[]WB+SvsKuQqRevxqMTKcO]Wuv$ή#%Z4ssbZhn vRߋ:畧'X+[,aźʑF,'8nsؖ豘! Rd[!cNbKP/Qr {wK黅Z>^ǠЅ:=OJ @"rKA)Ķ7&M#xЋל fe?DbRLZ(u 1fk!AR1XӬ^8btM6zcuN4\!6(* FP 9ࣉ RBj:tBZ71”h*Ztl*YCйFQV0F)P2a-ZZc99Qj,}.@"+ۤWt4I-K@PWeelTiVDƦlZ+I-:PYZN1BB hTavg}\˚!c!h+I^RtAi78:cULQ ta@5!?sUZXZcjY1HAŃPHJ1=J#^@ܖ}?˳!nj&{#m9 )YKR)Oʒ՟F(;5ٜ7BU jj.sG9EW\Lf&ͺ7(<%fAtv**;c.r#Gqu7GkcTmi:7$~JR\)H%jBq*!]R5j P[ V IlsPd:@P[ldTcʘcdbY t tfB+nT.%a}[aCtLZ4մ,{)SN Qҽحٻ6,WaM0iv7]x?&նvd)oAۤ(2eѢ#vxcD>Uust5(:٠-;y9Rmtۮ8U& J$[PyU벫ԕ8YdzjYǚژ[ѕ ,& Y؎`YlXkF*ZD@`|푽kj $*Z@'U68.H1O= VM8L6`Ԭ4 3*0 ne>id'X( ꛐI:Ci*dL'B,l*!l+u z^X szeY r3Jր1ukPP,Lh=4F[(2#χPzQ欼ҷR,q{`&b1S|bDJpPgSO1dvrP`=2t/Q*ACmj((Sѝ E4GRF U BH( {/uU qj2 R+a+uYh^UDI)be+%zM4f^n!GAgHH&9%R,[%Eࠄj`dV."2Db0T{4 {xWP>\FФBufpAu/fĥ"D$' GcE( / Us?` ]/ews= ^ׂ7{iW f=f/C&}6}`-$ >:%@uP< *}Yd :!JrIJ2Ba1%OpHv9ok=/W(39՜ HDNW2^i C.nk%TB;뒄 Xutq"+P(v5$B{Ռ6!XX[vGJgcQ0%]!]c~;Xd+I@d(.fWbd@BUq@ơ"2ΪU%CʰE(3A68F)PNX)j7hVҽ,I£j4ZIP*HxޢZM/-zrD;,e4{-w$a:J` ئQ}AgE0UK mry;Xq9twVpv~>5:L'U P]C*ih$ti#ll% fuwG(8v%[Y:Zm5EZSR%8O)O ]j31i"GFзܫǎ[zYf1uA;(XfhTYڂDe{zEշ߬GŰ%6vł*'XW6b4knk!g0EE^ڿ* Q 2ݡPʢ#6:XL&ݞA,+*R`ȪTqQc@sjo6i]0s0ە݁ bJ3i^& &ci@Zv!; tsr{:-{ *W!*|ƛւ LFsE Do+ (|U8P(-l4EP+Bb)AH 8QT5]QzB%aIPrmJ@Q[1xx@A4Dk̦rܛri J.ȎYxЬ*"SJ,豈%a4QAygjp N}TQ~ԋA5g7KqlX.ֆY[xu5P5A-yy]iy_Pz|,:  ]t(},WwVfO+?no'y~wS|@Q\Gzn(m`$=y۰}@b> }@b> }@b> }@b> }@b> }@bt> :"N8j,> B|P:>cQ> }@b> }@b> }@b> }@b> }@b> }@:V.1)2Z1zq> }@V> }@b> }@b> }@b> }@b> }@b> }@bбtpaT5*9Jkt >w8%z/P?Қٔ`y}bv3.^k(CYDes'=t8:,/KY~ zT1C]5tf|J5ĩ;|zV)R5ݥI_{SAOOOnǫ ]9xG$Q/gWuϋ {[dVCNNOd !M& ǫA$gA~i_q4D˞Bt8}:NVovpv|AAOTr2V=u'&Ͽr,w&J!䳺"khv;y&aɳ%qlL'߶jic=ޙihyv4ed8Fg-ld $[ H@-ld $[ H@-ld $[ H@-ld $[ H@-ld $[ H@-ld $[ G@^Xz~](GCWNW#+렇re_tRm0+pnDtU~,tEh9t"1ҕB1xѨ+F:]Q'̈;= 1KVC+BiG5GDWظf,tEh_"CWCyKu7>.]6>]Rm(փ]EzbUBYCof'ᴹ}~2~J`EOV'LV4ZO<RYsMzOnVR(dRc **u5>ɾ:dzĸ0s^??@cMv7bwC<<;J9;avPfFST<3j聲Qm7nk"PZ܏LJvq4tEp ]P*tutE!mً-(O/[Ns K9F/Ixutͧo|Q^BR˵^E7n6T8Op4D{X=o?>?[˶Dh6ZDv9xiN,m*jl:F[F#݊љYRj{HD-Qz`%ǣ c;vCkގ1T`=""j4NWc+Ox󡅇NWAaX hQ:]JgRN2r4tEphnPjtG+%7J<wlq;µeGqjGt%OW2]}K}#+rI;vf4tEph h?tE(c:F Zw(tBmlRy4ו.w.?w] "6ECg_vbj_t룿Nߟ# Lᕫ:B?ɔ:L)N]Do_oO.+ח \~过OT\Y[{˳WS0]_>2pۛ7gcmy/>#>DD?jneuo6;uy#nvvzZ9%՜2*쳶)^68^XnI%My&?]_Ļyj{s:+70:¿_&eݙV*m*:VEom7wآDYBU+нzFk=˲~]iwYhq>[ V/͗a 7##۟ߜut rK`0Dfvqνbo8[tgû9"bc/gGts˷zŭN,W7oXBT@>swmI_H~0ָ$8`CG5E2|VW=CJT̗E#it}UwU f#. a>nܯZz~Oo_n8ⳬy ĮC ZS 0ث1 SFx+& @*Tlvsv+Qκ^]l{N4w4LW{{~Y?U$Տ+](~ lqD|I74r)q ǧ!%&_hO?0|d(0_N] Fi9,%ES{jq8=oh^xi+Am=Q6"P[hmI^3[㙽+p>|`FKЕO #|*Y;'HQk=pA%z|RE#)/sZ8=>%Uf#frZgp5bwˇy *sF:Vma돷|gݫ8{Q}CL ~Ӻqא#~"XjXqOǒ} udHdS8ԧL6>}vU͇_hCoV zk4d$Ǡ<&,3bE0JxOW8omLP"kGDL $NrT" @M$*\"0.&nr̢wOFntJƯ1w٠Z̴3a#9_VdRּml4m\n'"X%N}n8Ih*Q!t1"C BˋiC O $POJ8)1y-*yE&qᨣQd F"%E*(4EcRLrrcwyJ1p.^+&n`;X1|7E&vj>V[[f0Mzq-5|y0#LrRN+#=بARRZFdv;=|m B>(:<O2z$QG(.},۳;)<0rucuqk"G$Xϛ)*}2IpvRy$$jp{/T F7zZl}G .DJ{v}[@lBK%@<+ѧ]z: oa#`0X?tKݽ@ch'>4IhJ},tst>ڀvGp.zd0 ʩZmkdoW}?)w^sddž@p_~k~B}{̿Wf: &Wʫ)39pzT(&A#6GƻT=ù )xk1!5beg+1H8K*4;$5D#e?AcdHw(({X<*%\FŎAQ#;ji UFFTnVsʭHw\g/?^{Sil[&)E֠a8ڨ6+T0̿㐸?ǽEhOͼŋ_k7>!n8?ܸ9M.b~zaoٻǟ#E#$7|IX~bcË^n\&9_ͽ1cqx53צZ\1?QeƁUzBʏ&BƣaDaNΡ4uN~Qd4 G~1~MLۈH%ͯ?Ҝ?~j6]zU颺z#/`TZX"j̆BL{T*. lAV2OVc#6I|GD2M=Mn>\uN3+;Ns -r {_dmI slfdz3u\^Ez>ϦGNjNJݽctGI`<[O=d)V`g&V˒VQ^J-}Ɉ֑BJ*t*Ӂy0|azE@/.)Gߝ=}ܢuqoњh}?mL1_x\uA^xXuʙj{ilD9Z*#>Wbj-"̦ j3_GOFh]9ԣA%^=<G(R>/XD%PGb4:ɬ,s"URymvvc\zuK}FU6NC9;D>0UG:v}ҁZZD`< ,mPs82h%TIn<ōй^0dP F}j효g m6^}n~7dׅ0nqdoټwEDwcqInO7ۤfWzo]l4~uvں9uH;^gPdlA.EG蛲Sۛ t&ye"W$\,BR2QlJDZF 0+ l~px_fZYy18-(K$FEI]rgW3UxY8{MlSڮ`>>j,@GRԔr|\pWg)Ȥe.٥bRTy $hPTQAmSl;KXG%ьKn@ydZ +j1qv+*\N$V7A=9`b`:]8!MNx%׷= I.w5=5~ް/<ΒGIChpOH!DT8I2OMN ASs- tEFet1yO2 2m:&YtMJi~Q*KR8W)fơXh:,Vlyf6ۋ22=tsӛi mP`~ПξrV2&]T‚&BxБDU"H|T$^ rCv\d{8˰ɕJx%fvPhaU HЦ)tٍ~4+.橠vq*m:~RҊ#dBVp hى`X>eJB c ]e3To$!#3A#D A1`&($Q1H:ՠ+ٍS x*Xl~0";DzNq2Ũ?$bK%!Q՝iϤe(-7)8Iid ZqGJ Li*Bp+ٍM427Yl5 5 :\Vpр&H=7E5[;KRF}A&50TACd*.O͎CCzvnk56rj fh( =ǧ~R<1%h..SDͻ >|.1*)_;4.u^yNkh^Τ?R-B3?È+4j1$NEJ84 01(|%4LR"MF"L  9\謏C28ᔕ@T<&ХO $g%g[ypidnyo]v MY%1_}"ʸwً1=ˤU3_F@{"5y Ϛ(gm^'w "M\&7qt_P߾Q ~e.s'%Ww>P/(YK0.e=k=qn?fg-H~0=: ]4O7m|;,-S~].ڽǑ>Wz~|y{Myoqu0E-V~g6 -ՋW/˼=>J{$ƨ ]F:?p6V7VdqyR-_}|ī7Ԗ3źI u$.8fj番1|0cfq}&?㦞;F}~ghц =Bsvih-sT i4.+g RD;=;9Sġie,b$bD|$@KJATLQiȆ8qLYBq"҄%,3%\D)L2(gya=+&Ξz_-ϱ2QaA25A9%Bc8>X*-k'U`, Tw__8_P(3)oyia0&P#zSalvB-Z[4 c xq 5&$e == Q& oܚ|)ah =p Bp[HaȕuT + $D_t+sxcZe=AsIRmC:0/T,%%#6o \N=qO2)w |4ĸYA6*fB6*Z\$~az6ضY"AڤxlCZPo]ZC腒yoaIU&޳/'X;hM ɤ+w?} Qke"(rRJq&4/Yp GNDn#ϣ0NzK" /˒:$xvɶ{ %8,2S<㻸ɳgp$3!,Nk%S`MZhyKl(FBz:my+QhJ7&WS&+8W+Z-m߬IQJdM.} {hu2GwV:TIxfkYM TR5Jg%6{O ^J̀QI]GTBww]zdF xuuEոrGt:ms|YwH-+j9)3|w-gIDs->&o"y3T}+t~OѥAӸ_rӣ9qy ;WY ?vR9:|=a`j_7w# E^{P8}6t@{+e@XBUr:Q8tNwt 9Hk8DOlM,Q(J*^$)*y4չ86tw˨H@h )Kn|L.˨rD"7sٖuYRwzנJ{tx8fo[b|;?8n3'?d~~x-`26Z9 cRHAgȑIn"@K wV%˹q$A,g>qM&'ǵ*v>o<.{^'fnko)ƪb,פYrDɞ_UѮR$2٤ 8ЛDyULj '7>/g]ŤSNpK:i8\0V"΃@s 8 \P&֚ą`Q !LJL*&V-vŶYRRM +{ j&sb:+@!lΚX恔H 20REH9 [2I1JkS,Kw_8b|Ӄ*HԸŧܬJT1euA6.k40@&) 1C>Plqq :g88> .sy= 2QoЁ:ť+=ԔfXQ.;ŚAUl0Y)iq$`ަQ@V)h ,eGZF$F<ȲceNղ5(Cc`b-cH> s=I@\:M5`b.|*|9SwYgjg30> r3ߛ~5[hJ bO _jHa{#k8Uf`v7Zq[?$k'3*aRA θdb~JJ.î҉?VGzA::%0h$#xH'Qa[Bژ> U,\|?,h6Ѩfᨂ=dF]4W.QGŸ&8RF:> C|z+dc`lNAѩ3Ǹu/~7"G_z_=÷o'7ZJ"Veaqsĸ^Z)~>_Bu9bP‹ ~FOIW:* r?ICwhX,ve ?yѪiC欛c̸㢁{'&!F/)Y,/X(huz|;{CYH| eoNRXB:ĔVZ+9xA&5H&yv:s&e#O †i"xI7. \ >FHBTb"pyIrQu[;_Il}Mk.;Uzv 4awe1E0v~Ka0elc7|S}3)Mv~'oy8G?r-ˀO,58WWg\_|IަՠTsE 5AVFxV3+L}S}v^4.1E } jJHnЦU7mkey1ӻY$kPټ{_Ԭ7|ydFMN?}{cs/#=w]E4.IZJ;j`w翏z|lImpfγv˖T$dWH-#!rcJ0䁖3 KKQU~ub :Ϋi#m{tVJt$dbT땭H- 9Ԇ2مbBK#)2# Cĵ[0?'ƫp3q,t0iKQ2^Rpc+F4 pgc$Ra<Ũñ ±aetp:^ikziq1X(PRMOfR 3ŸB`U%Ԣ rѦ*iHֈw髞ygh%<;lՏ.vz+E}2ϨJb_Ժ9+ejS8O?m>Eew;ˉiu)|,-G͎.P_d1F*%I:SS%Qgz8_~!vd D *qEмXQMHcivWTW}U]]`ˤG^~_@yDieKisσIYRo@.pPR1,! bAIMTQ ^`B<UXaj3jXd2b=6MVHKDyZ#g{`Ǔk'Z}S=iyR s;,Lcf|jWtֶz Mop'bLRB.]/]Y'Z.MNҦ UbPSjzd_ӵ­'rhCju^=,g1}U8լ:9ܺ^w0>yn~OwW\}=|r뮃y9@F+./d*'` 'نSs%>Wظ\02Egca{Epna7kQ4RN'!=j=Nf8`Ъb٢LN9ޜ4wo<~^qdshݶBa.685<9o7ծm$h%b Ai/\Nb$QD%'ȢGk,NcDYE 6AsglrD`p'M2mWhwVdǦƫgovu.qMQvEgYȲERȧѷeoBMVD"ԿƐO!؈4jp~!^A8rUș1sn7$Á:SIZ:HWQQ"b֔X6xd*#)M92+=ckfU^>/Zg~ykx-z4sP:EQэ, JGcu! iBCQR ԘgWZ%t;"E Rr]+Q* J3U@UAl -zvPݱP^Wkfݫ {omީJn\','/I/ >>07y9*NgD`%kgv>EPβ:;S(| =M.irްHcD01:9RL݉NÕݜLg NHg@!2a !Vr #lEu") qP(f-7qZJ_KwnHբzm,[WI~76?a Iy:c>q{!ߛg'bY펪6 '.@}` [,5EnNEKb&0G HF-lqeq4<זn8G,~Ϫj<~PHvHoi4@5,aP|LxYn =]9Κ'Q =rM6SQFȰ#i`}9.OΧ!+cFବ+/~);:ΆE8s?@ ;Ïw߽`:h ̂pDg& ز|+WW¸ZZJϲ+i].)JäF.|_V)ݢeX{ :oCxˡb{ P^>60w劭ՃQ,@LWrT~&6V}-g@{{6ϧ<1r~9';GwH;3nVuQ2@n;9{E[ ! '˹JgV `y"q@u4OXy ~פ 9=) N)|IN!  ,'"@ Y^nC~k_㽜[ Թ uǖ,|'pZ]&-ν]bT]}Gn4 Lth vgmeշΚ.S|*QOKMՆc"`k U cJ#68R$#?g)Oc.*5X@U=6~֑#rۡo{~O\9Jېe(E6vIe^F:<ɲ߆E |ozX\uphei=\I?Z6-`4Sxlv׸OѤ5k@ VEVPYvH0HK~Xǯ{@ח]D:Xa_7=zc8Av@PFKI[@`e DH +J9c"\x iArl^":oggCfz<[OlJٝ/ ^!כ\Aj( JK ^jȰ`ۇ/S!}8^+nj6ke-e` 0GM,+&HV;||]qv>rz8WgQmgȄrb^UxXwKlT/aN(Esfͭ$G~ZՖΌyxJ̑W/왵_'Ɂym!AHHhSxfrB(S L0Sd*j(>0b֑aSA[R.k%%RHD#i95rF)w1XybbvN<g(κa3؃[EO+0&I|1U/$msNAjDn9 iY( UG^-}jqDQT4 چ\9 8&1zmXAB̃bteA [ PَR;(/`y@.6~tr/27X蝍T +5(pM4GAw kn:0911fA}l;or*6{֦o娈ɧq^$^@,֧J$!n G.uX#XJC2:I4v '"3!ib’H RPL!fe.U%T8 `ra20p }-p(a>6h=B\.;Q|w$V ݤL<Ta><> }~sP$:j4X0Nq(TYP*`)v,(e*O;.|'zEO'y4CB։ }V @Py,'3B$O+'uF+D=;xP@pц Yk1UY81!ɩ0"g0ln, Bgb6oĢ雿iy";[Ǐ_^TOS8a*hF L*h e] (ܤxŪXt5,q(*{&6Bl#dgvJDJbg;a]·Ό SNfXiK\܊ʼnAUӳSAVv,3ާISg.?:(WK^ \=QK#TcT_J!5u)*I ︿POIEI_#rC?CW_ rr޻_lAy Co_;J>D3ΒJX~up?u⤜=]A+6i8ː]` 2g0'`B,z}ax~a^ȸ,<\E?JdHOYQ}1|<I8 zѪp|PN'>lw B dQ~utrx"LBO.|>7d%s: 3-H0;`DCI]PէlC*Y>._^AUDZlrmLH+P 0EX~R+ezv&Y;d⤼RlS XNdKO{y;``;ؙ~4xu%d'^ê,ۑ-K]+*<CKvb6jg;p+ػrz!4b%M[ߵ~6<> ]4ok"ӭ.]힀&܅ۏCʽm@ZKM#Q}˷ǓyRSn6nd~6ޫ]]nw}bcj&7M ۂV[ U?kkTKw_UCkK%X7T7`orӬx!>R\xOZ\u{ɓg@~r=̵nn.h=O p>?Rl7>H2]ޫX{v۷X{Mh({ꞎ{}m;uv5ǭ]7 OwC,+4G2Tc"L%r< Bdȥ j¯r˩:bs9wQw 8\|5I#X*yfIie; &R萸NeGЫze|,Wm7ܸ˶܍-{'켭8_pX5|@T¾e(@wS/^(3&l`ϕj?jO45ZZ!*}f.o(Ɓ+ /4730`iń~;ZB/rGYꧣPJH1oX&3Z- axG]0[LL1)χd o.V:P|*۵̒rNEРa1`L):E2ZjIHY`\FiǨj:\,=27І{U}ykV۩-2aLeeS.ĦKjMJd2iGl1>,f)&ʁ'Lvp ̀ J[1۱@ņ݊H/E8).b![s% Fª:y1:!d P2?Z)BDWg&Iӧ6m%1ERC-92&`耢v^&/'9mpP=!yPӒ6r҆Nf͗K7i˯7TUCXB{7ps1 \eosN' 8GV<[MslP*i%SFd&TuLf{SjlQ );ԠB$~)q}to',h}nyr9$o$>w]!A.#;z!v$dƠ~_}r2(ID5ș)Y#9˒#ȍt(H‹l2p t%*!V-O˸8__mѺكħ#̤L:Q59hE)*'.IUdAY)=sVQޞ}#M4Ѱ`狅|`sްroG}:4b1;MK˦iST ls?qCrRoWQƃM |PdFCgC6,[A`Ƣt9mײ~:Mge?.VK{<#m?Mm]cN.辶 1&btñ{wθ:|qErc}(;wrZۼռyMzT8NߺArqNŕwjJtS׿un ?RzV]=t2Q0a(=صAH[2ӫ(3W qFR r>?'O+풍']sNId$I֖W)`z֮S7Y򝃣|s#eEE J|yZt ãk_֛lpF{ɫݒDеş%Vu4k4~Ո?ZVoj!wrW'"o..EGÏ?L݋ˮ$꡷}VKMh6NfeϘpy>IcGΑWL.OaOb79i[{?,+Xں(M4ʗ#7"FKirxV1L'kN? ڣeV*8Ygǔ2w̔QZ!yk `JL K%Af6>J%.2..{g9W>Ш&ΖY[!%rXUhx(-rԜ1>R!"I.L2YDO4A yP3rޙWŧR1m9dg#3)&'IyR-$HckLSYK-eڊMW鼄+ՠf͓]ˌT)S|v JzwLq1.>X!m-XN ǩwUQa-]g$WO[@d1N,z. $,b޳r@Ɛ\\B@l.Er:ER?82Tmd&XTj3cW,4#*,\IT0 9"_ ;=}9=Y\|㈍`L"*4SAGh-'Z ] q-̊^C1\Lz/()QTU"ێ ,2 TFj<~6k&PP jG{ߌ' 9k(d*gZ$Lfx%ceUVr.x\L̐+L&&P$aq$/" dT'j¨&xKCAjcWDю8"}3%+UjR UD= p9/Ih/%CuXC6N$51D.8S*0!2P \[EDC M\C6KvU5Jjo7ş%&Ƚ4QXQe|<-(@' x[,q)pq06DZ\v*_q΍ nt~!gmzLսۥqvlB~;Kq4e統:&Ye'֍OSa<iT J}dZ]L ţ(;e,$H2MŚu tvFnQ3Rx ${!K}S[qat"\@gt.)Î{gKζ<]?b.§-6qIa~ױu g7 LyNbF2|ՂR"0« .,jZ \Xb .f ŖC/1`ՋգF/ WWIkգ #Jps lAZpU}iv8i:\qDWo\^k K?NÚ hJpɿ23gݭ&Y 4\&L <N0qǻ?6+./NNdl,4om@Gоq2MS̊{=>9Ϸ,-Ef~O'08ӣ|n-%dן-xU9=Rs3(oFJP Z4J+޸hSS(weq\9)>OgeYQuX<2 ȍK>6Lr<0*[Vl>;=MC}#|݊$sugi~0Ys@UcbTt>k*0d DU^I&A2qbZ*{ڇo7(zoo.>gw|Dw{rnޥwm7s2/6͡fln{tԓIJg_:xHҏmb`)0hlRs_&:,pOSNzءDrpD|@DV(Q1Ԇ0jKCNV*JiݱuZYi삱N:*UAq%F71 Қ,J gl )3^JH2&m#GEo@,~3$\|V|}e+1&3ڬy-eopYl Hiz62<Ho%"g&(RW*URIF+ı"PEDۀ+@q@I(XN|*MnGMHd,Qr&fia9Kc\J`E π+6\7_dgښ,2`] $%A[4! 73*5F1$ƺ}tPB:@%&$:YbkD ŊfNH6'" Suǒ}/:[xgߛ+7H9-/p%>Eo7! \jF(O[եr;/ ܁(?. 1T~w-H٢|m?.~V͛Lկ |bE1q>L|VBG A,_|-m?>q3N.aT co홪ֽy|,]T?ŻU`@URk?L|lHo8z3+.A Iڞu]7Ifu|EpT'FQ%BVBHBhfi~0hzl BtUiA9׾Ft>u*c E\ DSM|:#Z=.4טV*.u "9m(ɫ Ֆrw,S\;56ZD4rmA[K/vgP|!wDJZw.^J!q'cLk0 O s+HsFHSA16Oq{mu6vZSy pf4L7uoVM7:Xܧx(s5Q4k1Y˴ 1\TygLDlMs|BB1MJX)1!u2c#Jhxkۥ@K]n:EVvoH)3Cj#څgL!>kNզ#),f [tIOP Sx[g l]`sbzxte=nhe-j6kxHo_xm y0v[|;R05 wbοGpUq m5}Aa˩xEӦgbb.1W4NN쉓C&̭/Ai?W+۵$>}e?!'3 |͂kZ掻wܽ݁\A3B8\Fj5KL0(!ěȐ'#!@.ZDbܝrX39U\$Be25ZN#A"=&bZӐOZd[@q>O|5CFhP؛ռ NdN.Bo[h"9%!w&Hq(n0%h*# |@MFF `#C jkThA)JJضɷg}8Twj٫]_ss9|!fI 0vSŻAB8l?SnGBт+ &QYbTB1c0ɝĞMΞׯ֗Ro7 ӽqyģVy:1!2kI@.)cr.<%U0Zks P%QD[v~XX:;ºE~_ĤA&N9|>:BS^Gn gyL ;qCq JX,֬YG o'=qt)WBsŽ'kWyrZR`l cG ?sxRRkK@DTVpզ:%Xu`Ggm1SW&?k (%} AE$׭Cp"9~0$Z6_k=8ȂK/ -4W.v@CNs/{;P1"`)\0^PcU%"d:!rJ  e]7|\yƟeͥf;M˓~sE UpT($ e1zOd4Z~]SX`sP<1C*& \S>RXk$!&^k_KU9S+G@ձ)Sv4sdx[s+zx['dWtcQV9KU񖎻w Y+Jj Y]jZgm =֨%%Z+mո-e4-(G\Bթх(|[;N9N6 ص!Ǡ}7"$]bZB*2Mu%V̺MHj .x0wncM锎anynw69'U$y|?cN  L3ţ p{"B28O sebJ ̫""乼y3 ax>@<#.In'D$%tb܂JKJ="5N5ÊV+.lN~GYvisrzrRڪv̆'I&Ff,:1&HK(!.SJJ`q3<dv:Skz:Y 9/eHx8x̥7H(0N[KETpJCZ|NA:px2D!"n+%.qJAY!' *Z`qVc|/:YMG!}J= nB.͕l /a=TՅu3.LT&d߄'̺\L)M)z8S3DnV&AٙǛz֋o+ 5ܑB ]dC/MF׽\_ rVq9:hr'dl4˱) ~ZL 9˧ D`|}2opVF6íf=ZLj I7Ŀtߍ.t.$͞tlc@>I~C4Cm}P?~ߌMԉT?_rpDwÚ^7>5rQO ur:ִV$o$n׉̫FېA%ʊ]Bb?x\FWmu^ i$8$񈑬ׅؼA5շߪX"UwHIssr*F)(K\D/*]@>tkjc&vzh4;j?)kQ+^8mw? Ͱ'!{J:nCi鑟t1v U RUcuQfUbmPw) ۄfw|Tk3ػ^"=wuDgZ{_iQ|,I^a[r$zE%YָGMêbT5rHwzH $9(*-ayA F"#ULR8&"I2o/cжY<~bsc)M6 u/v|=1l@ُݹ]y::+^I=q4ɀyx>TB;vWB\>G"X+"*@}5{Q3fF͜ &'㺌?d'赆|Yi!Pغ%kM'6DB-ME-7 D?uy9an6rD%ͣ{WMSutkk9<ݥ#4]{H`Ҡ|z?~wEJ &hy8)-z7kiTs? L[g6KZi= )*rJd e!}H 1q4wp%/ o oP#&8lj߿jD+-@ LuneGGZXZE OL3Ub9̩{GgG M쉭kr؃h YPQFGυV()5khu:% Rpc)j۞0[>[{sU|IoC.nGy8<}y|5!z|0X~_r}޶op4lϲ zmw^q^[ٝ[[لj5I !~חl:WK3Sz!?*k,k[6X{SH̄{M[QИΕo26p]աbI<Лv.ZA»Srp'YQIC ϙ3$EJDg ,/ƈ4NScpu9C"e`ThhR5rLgCfݥd;S}tKʙđGcc,")2eHJ笎I2K$)Q#xfwAo!Z橏PIk (N)}$RWXkhq>8i9|ܕtbdttU8]0~]A>Js>ꮇM=I[TOfSea^7b]#$D8N4dZ^Q27 7EQjBHΑZ\!l6 q,tJ`U ~R*&XLXV)fƮX{,<),K-b,7;yi/7Rh퓰AiWe0"@\b/qc]Q'א=12lr'!(fИBߎh F&JFbGl7+]lvtuڽjƣ%))ɼЃ#E։h0ExҞ&nBf, ,FRYﴠ2ZЋ<*Ě,2%Qu!@<:QY[k~s+Xl~슈0"{D\5E2ɨ6 ?^AT(N< )DŽ0"opjTh(`) 2(Q@q(VAUġvQ\bfN8g=*(QpQG:ђ_Qe RZ 2"h P挔q!p+xXlv슇0atȵ?av&N ~cm+2/LT9ZuS$}!;y|Gҡ|G҉|D 4F@ g@A'Ƹh8zȅf" B+OXTY:Yfy+jG#u ol,WA [sHfi-?~?D.[jl4Gy^=/Ӳ2[eU1@OM٣zgqQg-r1\hpx:B[Ί}~a?i:Ii^)zdp.paB]zVKW[fg}xԈ3A^]5⵷4tz~oHټknvuƆ)ڽb0V? /f f5:hx8i`_4\#c{4H{Kgy8ǃg-:R+@2'䤎tQ;CA1YCrzStQ1q68G!ǟp}N(yKsW%y£jpm|ʳ|E3S) w@Pb :($A" r1dbIC(NH5劰0B[_zϒ@)t1*-X Y1q6ԳZlV nDK-5!^Z8N!8*p) EqId,S^J?? p(z^38HBPBAC2A]MbO+oYZBx @P_MCܲH!X" bGO{遣pt@h9-^M&tZ $\J7w衹5Do?N1/2*;R0؋p}[w}0F_ _*>9N&PŁygO>ͫ]ESMQ"3C0Mg{ %̏_uFD)~1&{c9<}N7=2ӛi f(_~ *ﰸ**$TF ̊hX + vr>IP& *-rqugj,2RqsB0 =S8YZ]77?)I8!Ӂd(-=4p䴇WpjNi_fB \eq9Қ> \I"8%B =*ĩUvOpWhO)v,.S+dqpJS'yBp~\eqddiӻWr$g(r:,8YC,ہ+pP82\m'83>0WIiD +jסĮX \eq%?Ji(#U>C0?FϬz6~8MN&g2G_&S^ uTļ`t(qv_N>cG]dbP݉mӑ.LIs_WhjLݽ^HT7 Zb7'_jf@skڭrc H\g|.hCL$]TzgLD8qkTzcmj.=*%ؒti Kq|rƀ}2>yWOR0]ɳ?OΘfu Ws LSJ E9Tn=g36dlƣبB"N茔 -y :c((ɖJp@J@Y%A$+3fgɁ֜ˠ#ʁu3rv7 :e{{y6ckWWY!Ur)EV Z$<G*) +M"-sE'Tқx" +v(K8ڋgDYxPP֞CޙڏZ< {?#sw,;u[ rNeβp_u 9wN-dykRpt `$D7In Шx=Vk"=K㴸gѨ[º CֈgRH"{}N)Z$P<4P30VG}宱:v *|d608N0 DoS̒r$,+j<,Dt,HmsޡŃ,;f]QѫSmMz110KX57KE@Ù7A򸔛6{=f _ /:;#]QdǴ_x[VWBg\eN܍#/iqʨD9R`"'_ GCgyp'|%(Ɓ`hʎδD8ry%)4U_X*GYdɁ %K6i#\[i߆I?-uy8?牋)6HEI׃H=~m/j櫵x|BK(Zr9GxDF,zYΉs:>ƌZӟ=X /5vQW {`it8g狽foל#demo4k[iQcdJ(7tѶ o[F1ˬ3'ژƓZ?],d<^nfhUӺ*gW]>d۪m B]:/$q2Ci0Ƀ6?ߟI.&_pzNRG?|O??}8?>7O e %O& ذF*j3 +e\o\a> }|qr1-ٵ/o)aa9 .'Xb-6hҴriaputz vY|u4xs/^0z"udKf|aV|`iv{+9Fґ[ hb`bJLh%s`M1K;ʥCr8, p睎,7q+JQlT@ B(C`%`}[<O3 ,lyOSnv>G<sKy}7KLxyG<6e4 48S2"Ȭgý^=t \i:QB"Q&g;ΕxAoM^RGR( s#mtdL`1m)YJHzCǏ>'6Xw]UF2sRdOso)2 .Ae;hT-W}aXN`LdPj iqg")^SZ2v@26K{Y{Y|;96X[+/fgP|%w ׌7v2rUߗPT }]Aqeٞ|[K=|slM}bK~;:E侼^wg+i- ء5X*qK-EF ohww.mu.8Z@@-WJl2gVapٗ T68͚䍳s{~Ο]*o3h{+y\$pzYXw^KU8`Laƈ wW,ߞj1Oiߺĺy O~K9`]d'8# Jy0)(F fKN/ KR6kk3%ge,k {?f"/:#a+:{R>ęmr3>s6,}g؃XoՍ|+>J]>qJw& pid9dcJqg"p#r$Xr8g.{, ^1 Id_{@K:pIy KauSP!Fԉ DJ^SւKNtx@Cpk?\g35f @sy6*ɒt6϶LhZ"jP-3Z07xenSN$gFdr-t^@֥RW<({[ĝliݍ?uU;soPVYKJ*Zr2ud[uDʬ&j,ӺD%ppy V>b'۸Bֶ]#{ϟ*mUIJz!˳D1#Oۢ"Kw1N37?y2(5B6S mKns\z~.bf1xU? i![(jgas|Odlv3I aeԖ jF=ޫZrLFcnXNM̼@둻޲7 Shed* {`:q)bO":6(nU3[.3#!(O+T" ;N n&+N^ϪjoZu=٩֪Oɻ.$?C)r@wD\(ړH֥,'.M2kYըgږ0e=;.uĥ#e=~ٍ:&pUk[MI(%$Va̺*dT!x"Q3qBV$_){EITEC)66vfpv/,EyM]]}z, !Є8!)_?vQ:|n_6샃u3SؘoSKVc3#f@E8lLśh5kʌښR ,y,'WD&d9RЛ& N4fq(ڑ Ņ-?Vec"iuG,q}E\/v ??}>?[,sT O]6ګh$P:9'TbC}9ѳk[ɪjTҦ&j v`] .1ƶ ݌fӴb k7cam;`oxt6eH#XTM`df(ʓFRfۮj+@Y (:lkRe6jQPAu67fp>$vTb<Flf2k̈ndđoY T0*Z[0iiDQ",MQm3 ƙP WabP3!vʬFe dFy#01#6s7#~~bU1u6xbq#^<|ȋ/m&cJ"("(سw!J9d,DD N"y1xXuʇ1(V{*}䝣UEGܘLZͳӳk}Da_7w .:t;vD QGbj̏`_c=q:g|$XEJ ZqЁʊD&'Ӷiݥ%aTADfRQ.$|H^;Vng_=/S $?l4\ۇo<,2 /pm˧V۸갅z5׫WWD̨WIZiWSԭ֖=TBQ;Sj1v>hݩRZZN1(sMv0WWew,Ѫa6#t9HQa: ;Ѱ6*Χٰz,uˆ$:yJl=Dj;U1"ǤL.A۬f}I[0%GB@X%'#19z4[Q9+|G`z;SvvK>c<;as}Mf}yjxq}&S$r"Ja"RD8Ld6F$Bj@syf;#`&</դAAsSGRqBt3)GDnjRj4Fh_AvAH-lr ђ/p{s 1vT4Cee_ EX?ߝ<Χt)k9ȜD8$$&ϓlB&F_~9CG6`4Ÿ.Ke\0{|)AZH["'^ۧ׭v?ڔ:Zoa+YNiwu:z&u$ |4utQ^kio8ʅP1v4\WwZ4WvFq4Gdms/^2 56М&JQwN¼m̉= V n ]+wn ;,(oV9tD#,"'A,{qq(BԠ%}1N~t&λVoI@~>hؑf}Z|dpwsE^rnзc|L S1dF1p% aE5pu +J cHJV1y Ĭwݯ?7rj0N6Va +Vq2lqYetqѵ9Nuw*~᝭6dIKdQQLtR ;U-J2t),:s֨vƕ"]AWv27e\,uޮ' GTB\~<,?3'L&Ws_{=eWӯtGPwZ e7?5g/ywO|`sA9gvIWyOY_V=`zr&L&2;gEy"SЀPdZLdZZ3ȔڭGV F.+@Zl`&vtU*}tUQq %ҕEr+, ]U4h:vbHW/*T0Xk ]UV>`TrhϦ@@NWtu?DKWCkҊ+]HW6@ Uá+ ] GOWRҕ338ur5izmzO[3, NRo@M䗿d>[NM5ϹqP';ao<HWޤM3`!`hU0h5;MWF~4M>.{^' ?Ma"c'cB5ɏtjxy2杙̟NgY\\,os| N$'=o^J21gmiC[g-Ncu5>هvI($|)R[p" ȡE<ƊHW/j/p@tU1p`誖G*J^"]iMO;PpuZ`T4K+t&CWRCUEihHWHrCRW X5b0ꪢET@pz{"c.#9{4rSgϦl{/9~\'~oݺiEWN.  @췏=o&:j *EddE]6ާ ڍtv9<nI?/ߤ:x܃z~p{94t˟_Ӽ\|io&9.'߼ȫׯn՚N~9=䕄PD sU!o~쨷FR>ޟW=cO&U6eWnzYƯ/`r?QҚ-{an9-l]V[mg{):ZO债 x1u tʜ٧.hN ƅVMZReQyX"+3`j0?5C+ZWc#~n_(kimΦe_ԁQ%E\a]Y.ʤܫ9͟ ~@<ܢY7 {iߠ}zIhQR EY-* !hOVzvh#9لشdfgY(aCʾdBNK2{ t}׸u38wWh|#=?핥lncO}jcK_cUAdgYbWV5<yOAB:(}H(bBI >hFFxF6Fbxb(DP4zBkWBgt2hlQ",Z`%"٦ cJh2\" VQ9%]&œţQ6&fp&iHady}hWӂ||{Nv fߦznD|[?~>O|uR^|ē,rVr+BvI{`Qb  &DR1|p!ӆxf4RN{ce9 Q }NՄ5B`=0JII(eJ#4a*}[n1{qrK]8rdFń\2N)GcU@[P͑OeͿ >];/g&ވr6=[gN~&e>{?yg{ˏ5;>}/^>gR_?+9-pRg VYh)hO*E8uU_?>in>m"y$9ŸƮ~`K[$[%_8twUҢ֫ z"Jֈղh{ {jf8ެ6T9fxw]M;\bݡrY\|6!Xu:k^CqK714s;Cd0هn㊡qW{!fа26|y3> :s3Uv쒭r]>m}xh iȄyē >bf< 5?] /y3 9ϗ(]'b2T_:Qq[~*ti߇"q.%s)'a̿NtŲ{,W:^,_9r2,fs7ETti=c>t3.?|{}3/λt~>kަÔw=b4~J m;m,;D "1Fnܬswov<^\hCBTJ;a#HـR1 %Jz4 /N)cH|}C,gUP}woAֿ[^ju:|7/oJYQMBSPFs *1B%VLBL*1&F\u Z^]> Rdž͍[n80}} l{)o OV]J::?vտ^mmj'.X.tU[W:]NuNVܹ6zraie}L&0xz9SfNz',UGҾXcPD C62'+c19[1h@v9jOn;/osXWEBS(z Q`)Dͺ VH% 0&7K]UmحQ )82 ܛǽNxԇy]_֔ޗo#5Tf:LY5E%%cgHǟ(y=&MGc.$N;kOμc,«>0rQ1E%Y<2Et#% "q]\6q^)4#YYC̵q[|>;?j>)‚YE Ɨ줓El"XK"e 6)oVUQyk52YM`|>M9ν(:ۑ߼Dx6~sL'CԏD)DpD[Cң{NcZܶK zswm"YLփEZ HuDE5*TAۥݗ]^=%[O4Ϊt1Yye~d'3LnrowÛC~n⧋Nj*}ed3qd dC>9Ba*kubݭmqމ~l?Bzz>=qW2 RENfCvYkeŝ+S˽u:WG>]ҵO3YkcfbMQR@KD37gO2h-sn۵E"*a)5 Y1"hXR *K66ܫM)cm#h:}BP ?ZAJHHQTM i3tR;й&59 q=Nl6l6K۞t=FEOP$CVWژ\PGR0Y2kTaaO6C3=W{7E,Vi}Ipc e?f?7Q{}_CH'W<[6XzoU[ꔷ-^~uv/f?}_ 4qIaWM-Ӥ\إ%E/Œ%l{=-qo{u/|u ׄ&Iciۉ3C>mJ(M{6}M{~ۃjnD\`6G'+;̿Y^/+{OMVWXhRM4ܺ7'<'׳VZ1vy>YC߿W LYx $Xw®'h@hUi[)5p8r erC1w6=pW5}m܄ݯclׂlm<1#~zjWYj>v+vT usH2CuHdmR¬IJauʝ6|S'v·#eޙ=nN PeR.JjSuF%D,|$FW(`.|lڮ_?[ւ2ښbKȧXLLUN ֹF9m I;/%Hu}U;#ji"ODhVΙCx꾓+ΆD;/D6I$ ˘.ibx2 MHD0+QFmNH54QԚ)4Hg}Pɳ_D`B Zt L$B,V{ Ԧ<ӚYJ'2d2HOP6gY K'rD bP8YW~j٧ʇ?.jt4ƤeS#gX=NA~dTL.jH2j )k05@KkkC}~u4̌B1dr{w@T[h ATv5uueZ͠V^NG4v19„ %9*JVPX23"8훛͉#T,=Gi!F梽PrIE_Q(Je#Y)gK#rڔ4j[v5}2-@MSS#}mP̤IXRb A[1CQFh[0vRš͐cq բ:õ.|JןmJ^BУ5d0&PVmZr6s+.Θw #EhH҅}5AՆc+X PL N)atMp% IZ[ƒC10n1)uF+cl_פA'DNKSh¬sI!ɘ.=l+q cvvDL޷:3>"nqٻvn6GtBJ-ڧnR>yv;Ν)N&/O Q(7!#;b+) @u1>YMnb 'Ne_a1M / B沙ހT=KKB8J2yVWK0v) tZRWyԱSұuoza#9$Tmc>2;=3'i@3!:3 ؜rm ="еo2ڪAAc8 .q€%a'+yOӖ5Fn|vIit,DPx_%oM.b 2XSZ>?Ya(Z 2$76څ6tDö2 kW^ TL2VZf(,t\`]ٻ6rcWJF}pN6U٤j>ypS$MROc7I 8#[A4>t7tpKU` 9a3Plh$ HHQ+NzU/Y&ξ M4s(C@s2Uqo= Xu:P C69[,X,]Ys0$r[kժGq^ ,3aXD:Ǵ/u'\9y7䠝SM ҸN#@R\n[x ̝(0 ~"Lj͋I\E"HK1nIE"G)#h|\`GRS_ikRQIQlsfiN(;ew W|cX؇̖]n{ӫfP˙:\d^Ho}l@띯#\ZDv69d;. rӲ8?^lM[ Y1 eWaM\o}(,m=MS_vAJwj3]z\SZ5MNѦ%vk~&qw.k,m%=m1#^oliŸג)E 0;y]./^ʪ#YՙU;ͪ{zQ QΤ:IsN||W.a,Y$=q`%ȐbŨcg,'7<7VlЭ~وWTe4]N@.pXO%XxIC&(PXZ'd&$Q @>LwZ D佖豉hj靁'jbSM"J5 o!hKdH}۞Id>kky QM:O̲*LlQl^p4-b + UIzG&B۫N~R3lXP:ES,zdG2ifQ!̡d>n5/ |v^敒a<noepe-.˯؛p=QzK&B5ݟ9ܲSrWu!#?ni_:DL՟Xrm3Z ݽ`" O}(>C4Sw|woՊ+ùլ'FH9՞puG-ǁ8NWi #z!Uxd!R&R/5eDDL #d Xy$RDkk>࿃ ;h2af[F|nnWe;)-[e|YeDãj^H287JNE(\3Yͤ|h5Z`(H4wFm6*GT KIwXTѡM+x qrZ:qާ(]~EYȳT~#Xi Dx$8 IB9\8VyCXX`Q 0jA#c3k!X'bԆm$T0Niʡ$6YX30#ݍhQ4iSMg/ZBtXd8iр/CszZ(\HlEjv]n)`u;"E RraQ贒6`Ѫ `H&H DO ܾW0Ԏ;fb|@qh$bHFƠTҀQ(DXi鹴D;zыE,Jn=2g/ Ќ5S擮NFƅ/i[  >_|kr1b&0o";f._.23)d8N*5XASg+)w}}]>I1I`gX#2`)"g#tv)33kv6^g i+s,9lB\:$UX+nAa q78{cҠ0&B>30BEtSL"Eb~ztz}C3W49A8;#vPtm>>qG\Qoٝ>JgeZ_/'?^f0{ K6_{]^cKrlW#`8fZz~Ró%׷t̪E`X`Ŵo,];y'7 NRp?Hje8vIȆ1Kqp6, R&T*@'ޗǙ덮@8͏?M}O_}͟||kq.mGKu ,_CK盯ޤ$М3_)J<.}zͪ8,`@3"E]MC{bV^ԠJ.oh/-w QC:,pd}PD0}QKo ^܈jտʺG-x>DafQ0E׆b`N^k X,@#^ӝ9Iu,,MV}jҥs6vik h Ul=,\w.;DO"?zbBRP?(r?`Hk3O]9v͡9Øh8oØhØh1De61L)ƺ@ e<(FgXvV`=V t~ XʶNt`oEQ E|{ Ժ =4>:D&X33%qhK>S5$I]>ӋgkK6B#ZWI\WIZ$JJkW/4"UX'[?vb Zk[2ͮ10nH' xY H(򹖜N)bxЄohc=Ϟ\0jX>qy 4ohQ(-KޔD([N6uT0TQ5 T$Bc9rE:rzGntw*=c)6-x;Pk20 fm$~It~A_ҥ)>6f" *vL̟ɧhhԳ3gz^*~d&|77Jqb)8kQH{ϒ%,IY@J]xg PO+r9[W .'-pҊɂJuy/x p{*K[s2Io]%)W8L:z9p%\I7@`Edk*Z &iE*I)uW/$Eo&= .-p2W %ǴWR=pe]%iEv@JXW/4H)՞KYk()@ZvlHQj>p*N:CT!uzK`j?qO}2DlTWN=fh"W .-p=6=*ItW/Er%q_Е|2F}M[$EHkP:*Q[P=E^R*;~(͙^!9ŢSױDtq/I ` τу{ ?Iw7hkۍ#Ii)o/އl֋1E^WPlli${٢4*"kV,1FH}?yDU{DMs2:凌]q8upί;vun{ȹͪCӫpU$t7 aa0v'w 9Og5D\^C?:~#ڍ#?zM7(_ׇgg{>C#śBT8~ mg#9eT{tn~8Zloz'a@:{'}>:ctG#fwGh_ejߚU~@AǏgj'nD+o' /-(Qfo}qDGL8E >x<|>-y i~8}H>o{vA\A߾ԋ[\իD}RRkP]h[&LѾ$*]u4|WK1=~{dㅛtOw|Wy~7wp]fd]Jz\vAɚB:+NS2F)77^[ ɔ\ݪJB*7e$Ki6kTsPz?8ٚ5S}:a}mͅ>TC@x+BZߞC&[ECI.ELt~N;"DK +FɖV1I;׵A3Z Ct%'hѵѦd2oj*  5Ts3: kJ:㍢+wFJ-pNC,yd0cF4Cb>olfV+Ө%gC {["Z#]r{}&6.M݃6hxe&dK*0̿Ko|v 4fƈa4By!* QN!yD!8<-@#Tr$e,qU 1Wm*YtT:DyKR\[%PJ.};oNB<i>PjH%5Væ tF=mOF8'֤iws=i:wrjsi g16c@FUy mBeI&$ՐR@jKIL0k /-@xCkmg͋jS-+5*dlzaCKUŽ9[0f %xDMU.Y{V+= RAv䛑$ZGviENeP4BZ%;6Ɂ)BAnQxl4JNwh A^AkaFhviɸl5 "ˮFPyU7JXɠ-|X]pl5M);:+B:̆N4<ƺgFCVjtwp擀IҼ1 [(9ԠMEX#6[ Ze3@XԠD$J(k! e 1P =(aL! #A [ ^c/ĥj*1K=IkB(}G ر3ĺmd:2<`X*f=f~ ;5͈ G 1=*4*8JPtd㫒9JٷUe ìcJv/6]NX H|@$> H|@$> H|@$> H|@$> H|@$> H|@߫,’|@ћ'>zgN|@OoՉH|@$> H|@$> H|@$> H|@$> H|@$> H|@0f8p]XѦGP&}@&$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@z> KK̺q̣u-Gʨ}@|$> H|@$> H|@$> H|@$> H|@$> H|@$> H|@z*>.OVճԔ[[\o.?, hw]WݯBd["Yv9%bc[mxݙؖm铡+b pZ˒њGOW)'HWOzAtuq)th}xt(zt=o ba9g2ܘBW@WWZ'HW'hӂaV/-=eu*^s `kh1tpbG *tt*-p+R*>vbN64~*qsJ^]}`ԷC̷C㢫t;:ZDWLCW] ]i6NW@zte,opKqrrX^}kfs;bdxw+x-<:5k]:!{[?lom3 3fJO~[I%MÇlr7ؿٞکvѾlņWT#y{دhu*c\~]m1?}X~ຝ濞osU0~M@z{yq6-]"6޵dW4O6P_7Ԅ`=̯ޟWfN/xtXi}=mk;eO3vj4Z[.'N5BG񠑍7Ң2SrZznΪNܺh2r>'kpXNpYf?v% ^I=5r_xA;+F{sz8]1$S+҂ ] +ۘzr2-tDWrAs[ң-(;Fn$ap/ö"O۽,`l~FIv2wփcV۲f&RgEĕNDhprm4NH"t\JHSUD.P; 6>\\u^RدS)]~eZLeV4Ͳ"e|e )OE44.o&Z,|6ϫZE ʦ" VC~SEaC W+#1YW,hpr}4@HhKޞ*109E\Yhb]Yޓ3gz 0PA?R'\"hrQ%^`4b2ıJ#NW^Y W$XG+Xmb\MgSW*1B~j@w -z  " 6\7ݗE"2|\JeNW)%MdK^o QA&U{ ^ݔ~."9GWLҫ1A=)rBZ&tSUJL4 kx@;T2qE~XW|վ\r\JmNWFH0HT>\\3jWR$\ @."\w'a,"6LĕdLvlYurm4+V|"SNW@pE  W,WEc]Z|UWgP bɅ~j@Sٙb+pܦ0"\`-U+ǂ+Vkubmp&Rh" 6˵2t\2WZ^_ѯDsԊ<0NY%ATҫ9蠒4r=@TcıgJ:7z+1y$Xxt "**$^VKH Xr՚q*C{ep& d]`ⱮH.h֘q*;ĕQ""\` Xm Dpu(@d]YDD+&\Z8'gq1B4b:u.t\J)3xh$㱮Xnv*]z3JLhr ɕgGSU?6ؕ+pܦW~rW浟ΓJz\IMgOU/ 2Xԁ+ו鄔Jx^ȱӇEYIHј,U,&!"tU:LS4 VV1b8;vJG;Tgx+VٱJ W'+k& W,x5jWo]JNWj%0"\`3Hr1բ W2ÅWzXW$WhpjU fx&Įz 60t B?rf@ST b\asF+.\\#bu"t\J/NWƾhU 9L-Rh>WsaRsx@K!:J߶/v6ֻvceM.MHo*ԥPyآnEn*U]ZWA*PZN(DnKAMNû w5w Z-iկO|yyǦ37H%_77=vldJD_f]J6-dy]T)R~xwO--o!xAsutQ# cTcg笼}.}mQp`r"~.V tpEƃ+Rk WRꄫzVDbhz7~j W2W[ /`  XXpjMb W'++pł}<"YjB m7 "ĮXxH6t\JLSĕ˘V`Ǯ㱮Xk^+ViR>zz Vrh몟\50E? % Wmz+# JD+.\ڡ S!q%"d]`-T4bc:XSĕ)!է2X6}`L[/tU4\jYec<0)F;3`x(M 2Bi&8~:&b0J%SMa\pEhpEr&Hj>H*=&qC" :@#˵JNFbR&\ sDb>U'$  X-+V)uu pEg hɱWJoNWN91:!bbU„+V)NWp1YW$0>\ZCp;́MojOA%շ{WʁC<4maz$\=#AXU: XwTv]քԜH[y,\\uY?&\$W:ś(mtlL(իβ`O Ȕ3yu=::v&F9A;ێzqYh:+k bKZ;_*_[}x>ܸIA\OД6+dUN;EK-Ȓ~Z_Z|g&/G{ Wm/Qh]O'WW.i tT^ts?X!Z/EKz:Ͽ|Hk'eθho~j˪s⮤r1s5𺶅-[i+m#ºB#Ѹ*/]n xS v:'D]\9P eUnLd\^4ύ%M"B)G3E`(pEU獪W(Mғ"uT .v`9/]_kNhi(2ϻ_.<^Q|\\oĶwM۟^fd~ſȨח=?x.v3էj 5s-t3a|v'GZdQW!tʾuq1I>e:D\ʶ=tZ]-3A[TT =^}Is_4ғswG7-шtjq?׫ Töը voz [c>'ٖ4= 7FOr494 "٣Ot _gqbkL}蟓"shEuwOb}56a<N> Z{;7 D MUrj'Ah(([ɵgWNMQWʗeGFF?t)ϣ/joK+Bɢ-ݫô{>{ʵ¾\W.>kH<}=}_6I~Ex@ Z}U:\2LռMRڮl|?mA˔ o u'+|6+wr--ə{ r{.ȯdur*S6oN܈ۥ۹5W:sߪoY5N|yꏤǦ (4B6fUÅj+^E03U}Ym~Xz6Y-_YW$3%l\Gm\'Kf'ʼ{t 7r6Ɋ,&G檧.)g庛e[#7ںE; t}qF<7rsBVolD҅jB &Xѕו05Xi@ i$QːV XoБ^ <'m*PόR{Xcy!eAWeKcje (s[B]@k zG- u>Z߸&;O"߷{fG gUY&5~OJe{=zQE/ο Ӯ6֗WRud>w.C:+jgARH~Nrsm?_9% Ka3x[g{ȍW!xpxl7Ȫ%Xigj_a4Qnf5YǯUEZV 6Ib"" 6Fw$sJ:)A!HB&Yt&щ@ lUH*f䬑~C!mSv]H3+~Έ6.,B'&&^yU$IeD#38!}~l }%m:'{5{˹37J+'o69e: C 3K=6i_h/d oc 0Y\5$#$RD!1 ϭ0^m{7r6#|I8x@WMG Hvht<ɝ[sarqҳ^io$Szzzw~6z?o[;:Y]oOX9܃vۭaLPOCAm "usOOaPkPd\clƪ [nڶ˄ޞ6}x|vQDmj'P,XkOlIRQ:jo)`!@ QmxD B2a6RgsƐTE F_$MJNQhYfNMÿ)&)袠BX4CĤҚ4DE\Jx[[e:,NS+%MQt,%4le ))@&x0="@$V/ԕ KiZ 'ӴƙO/%lMwϦ'}dQzu.f9Ω_UŸ-lcc8hĻ8Ve/U9%C !NaJkHY]@R>+(/!1H!75Z<#oTz&cf$GNzMj#eN^ n5b3rH\> ^zqlMf\^\u=j74oE/>Fu18_IpΥ JzR#UC/n>lCX#{;v6h w7QtIk$ ُϔhtycgJ\?1&dAc}ae\#!K^Y#iw||h{7||GEb+d'0FP D}0(dX;BAYh-D<#-.IEnŧ%ɌB("ShR=dV BKk݊u6UDo<;>]b+s^|w CE2:?5|/_N6W7z`:?a{>k?~aZ.kF 3:c(sB3ej R _#g__^?.@0!VRm:pwHOjZ,kl:ݰv7Z}et< ,X.h7O'Fhk| {eiwg889HX/OG /q~,bS_;xέE=~lH0Nٹ8cViަm^ϛz^osCZCmS=)ZSL)UgH ȳ[j+Tkrdض;7F+dH3BBD&SRlMfki/5jFuDCձE]/)Iu˜m$7|<5l5C]#<O=٬?ՃUƀ% J_q,BP )9'f͔3dS- $whElD lL]:(XĨP.Wɘ3mk[فDI+\d0 U*Qhe*B D>8Iuc9kFΚr_-OQSZ$ .em5*g֧Z QZ^e:e䁗ZRuDBQd=XꖗAL^ZQ !E5x&^eeZ͠ ~'wh0.,/YXPyQ'*؂32V͊T=9(Dqlxb-ݸ##u4젘`(Pr.Eh5 r e%dclɃ0mS׶[l!YJ@giT %={( E+/A<ډ mUft4n÷=U 騿օow%/_@У5An0=SnW"dLB 4ݽkg59fK[Y0)NAPrPza[HirRIɨUvdULD3%Q[SlGc4c.CWLDR)ڰ:SF )d-9$egCGY0 l]܌5;N6X7}~gYzFjw@"KJAtsI:zKFuɇLE}@):0Y|l/OԾ,:AyG{ $SB4b성D09nx?r+\yVЄ)ŵhkºVnMv{(Fs蒭ʅkN,4~MeЭ*N؊\y;8x~rM*߭o' Ggc@)n^.[WfA۽ڔ~9g;qPN/aT}7eѵwZ=^Ckki~MwW'İf-ort`>:^hr2;j5jrh7Rн#{H}èawir':ՊL.=x;w|ɾQUj'jYpԁ-ϟOP}>IU+wd^ɿ>}w<1OsOh߽~K_7geG5& ebTtn!q7Zx۫ ~w8ڗb)kR+Y[px^y80-L_^xjho9fh 6Xg\{qm0S!AKa0;Po_]G#bqɣɥZ.jde 1H$|=NtD5&#=ua\d<9dC*ʹړ>6X_@$,VYaOɞV6㉯$u/>EKd[`Րo86`%8qH-^e)'ly"ۺN2q:!g3țeFxoBf]ɗ1KmQAX01U:e6D< faw2uD%ݯәE0* 6*5#gsW!\ i[2 $I1,"j_.CXS80'8[޵q+ٿ2u;,H I\l|Z\|ZeIb{zX=3iiN`KtS$BOӀE(KH7TCM:yŸo$~i@ ђOxMFǒU,tFB(7:{LA]dzFS+6D[Pz0ZJB }ÑA”m0!C.u9^!]=)2&腀S$1J=D(hS(dz y&sbsݥ1GJ(/t I*Εf!t:1cAu`y21|s ,վbL!vRH" Y뢷<cWBaǦ!i_/?j8oM N>&nb:/nzf; eS4~X~6_>_5_}ͩ`к"Th;ԴStUརuJ`H+.W]g)dr- w/>w(w#߾zj{WUhn?eɝvH^?/~C,̍u#FJ_1WF4ǷM6YGMFĬ VO<]`UrH&_l(A$6gkr)I˙ ) fElUS%e{~-̹ h:V!v??iu|>l,b쵇z7duVMGrUT-Rrf7zᅷU{|E|qh[k..h.}Bo]fojo߾m n_K7oy>pU T׶f{C.pfav]~s›o^Y-|z{^\:7u/vˀVchq 7N>_Nm[5KVK-iYo=FuIQ 7,8YC oGncl~\)R Z Gys"5&Ԡ!1H@$1vc1v?H9<%IKR]Vg AV3^cw2d *(-ꂹ(/*YPQfƘ0QEclO3!~gʞoVb۩هZ讌W'2ts*}_v,^`upf%3[ EOZ{9ڂ%յڡ)W#8gѓw%d=X+FyT%g~Pg}㸆kK`9)dw\1F,Q'8ZfZq5W6?!RݑiAr)gc$ڼ{;dbgK{~CirO>gH>v9'~{C׬Ƕ6ƍ [vӼ__>7Î8Bbr+)RN`*!XҨIqL㇙d|ƐfyNЎ!́BICtKzỲ{gaK.ٺ LK &淄و~ [9`^#zfV%\Ȕ} /͛b^SO3@us5HD/u Rs[RD(?FYd,ؘ; BgͲYS.F~L&y}pC_8&~\}S5S+ӕ izMH4G}b}^SW 2'ڳ,^L/; OOaq٧dN&DVukէJO.Yȅ]s[Mڕ&n;EZj_OW;mb݀lUYFUlt=pl`0E+ZǮ+J~JX9]1`'`Ճ +ZNWʱB_"]V *` PSk骢Tpq5s2GݹyIͳ]Z[ ;\f۬"mO浙Ol~Zwᱼ.]Z"{8KTQQ1%L#7(7]߽Ct4ݬyګGs>\ݡ6 m=+ZpKB+%5 T `CDht1D O^>+RZBw4?k3R?[XejNS! ᠘0S qyd 4xB=!iG[ݵ)d5:pJ 2Bj!ɘ, &bE2>w@] Z*pcf`(SBHlqK.Ybh/}5֐Ǐ笨lc R[`ݜdљ!V΀> Q1ɦ}[&h܍EJl,1qۨe))c1L06YS?6ft(: ɲA+Xa*[ bkCH/!] #xoS ! 1sGʧRȖXWQpKAT fpaC I~Wg>UdGQ+i, lY .%|bmD1n{uTE"F&}.Vk`]IEQD1I8^ d6:}-Q|­mǃ =e'c:;~Ns[K#_yB8:'XaBdVcb$Ie}l!.b`٤, MHE@r"a'XO`@Qd=A6D)0YX"e6y1 ְ S'X҅זE [@$m]R~1d)e*+YZ9tL'X}`'4pg$YE̊2k 惚=Sb͡L k;r - e{5b%CNU۪6ڔ0b(n d<6"31tJ^٠*.nZѐ5*pp_'$nϐB$&3/LsT`ƺnж4٫)D]c(- R$mSa|NۜV3}>n'-J:PU2cFFX3)؀M0!Kl` qI'\0$bWx(*Ҕ[ 2KBc3KΤ,(bI_EKXv9ٻ6,W},w`f7aփF="7sz0vmv76I6OU:: 7CA 5q2Fl̷6Q],,xt0wD`L`)B`@YLq>J@ZAA6 L:A@gGJ-)!LEwV(J892rFj g (Ez@? }RAPSӣ J۱r=R"l.5ީ( EDhzd0GAU5_ӛ伖H 2oJ]MȲZ! aU#ǻCzM,dPgm0ȗ~9F{ۋ=Ril0ICQŘ@QE&i$% !pB6}fC< mg@ϯ>-zS ުA11C.0莼3  b:PT8=uw#tፍPMB)hzI+CH U20S )(yCs ~3XQ|E #.3(j4I(2ӊAU &ԊL75c[%$0AG*AP:S#fHA@8CEUwd!*T?yF*(@VT%gf H VEߓݿ?9~D ȓd} C4zc+tm i,!zԥvs;sP6D>G > j`5Ԙ jN!ns)5ڽIҎa<P9@1^ BٗjwXC-DShňCyp ( g@ @$B2БwqAkol:;% YBirL  "Q8TDYUPբ=`QySB/ DF )XNXjC$4]+it=iVԀ,'r(m@3HKފ>z2保%0lըχoAK Cb)M cs^ymNg]} Hxvq62I)S4 BztftrlѓйF4IvPp$rof1h֔]kM!J5# U@ &tB)*2jXQTl01Ru`\t!i#z8FQQc@sjo6i1s{a6+5iQc4A*Ez(tͤz՜ s@׺O^nY;^ CTP]xZ)EwѫxdmQiÛ  w:JNàe J F̀|1$zʤ躐Д n6E $wZ{*.(=uGm(}ₑ4 t Wh zu;TR5̸\"AC,gd^xЬ$cƛP6SJ,뤀+9a4gRAygJpG 6>zmbmj+7ڌ]hD1ү֕V/>_B*>A7]?@CoU^嗗GnKڙS̾XI }@b> }@b> }@b> }@b> }@bikmj 2Cd }@b> }@b> }@b> }@b> }@b> }@b>C׽[wP Ռh]I }@b> }@b> }@b> }@b> }@b> }@b\kbC> [XMӎ}@P> }@b> }@b> }@b> }@b> }@b> }@:nﳣՔZ=n[믷ڝk-ծY_],o"oC%lE3%y| b%Bۖ;) v"^BWv.KB%ҕh]].JV4pC V!a6CWWV0u"2] ]E% +C+t^ɩ\[fj> r{mj;^j z?tJ-EWq LWz0!`c3tVV hS+I&tutԺ%`\3tEp?i:x]Z?y"3]"]YR]qgtQ1i^]ctj܈conVr^]˳89nY2t7BZ!ۅm*oј4F w\QFw(y6[x ~w?׻:%vt_Q,ݠޚ.nAJ AՄY2z8Nkw&ETnb1 \\[Z?w&> p~-:rcOg_m=zcJﵧV =}lv9_Z0rN*pjk.pG7[khH Ѫ%Lp]3JF5u% VJV ѳ"8H.NW(+ko*.+M+tALWHW3 ke|+tEh:]JWqhǹ CSfjmmpS+LWHWՒhL3tEp]hBMauut6DWKpm3[z R;LW2aCoıc}#vɶkv%ڰPz9-Ô{3Q"+|3tEp}lV 1u"2] ])kZ~ܝhb~^o6#ou>{U~t԰?r^_#o79˳J'_tJmт [w%w=M*Й,b>v{2N-n(?b2 oS1PFvUR[+'jHGpkEZ3yG(dw/`TQ4穵[M-ڭ2k6Hpj ^y:MH6h[RE;fT%ķDiXU"]` jZ+BPtut壴16DW\;tEph{('=g4t+Gh/1(#UNX]h+¶BWVM~PZ.8t%8XĠ'`#w;jwYD95u%+tءWB'+&n7{};ZLWHW֬?p1?hLXoݞ= S+j[Ee, (o7S뎅G=oVݐ&ʝf0uaRP‡(Q5DWU;; DkP:tutep%uEmnz1FB)ҕ*}Ovpmm|1H(g:@ҊB|;Rf* ]ZNW0]"] ]Ү]3tJV*3u"*b7vEn׋f֮R;du+WB)OW9vpÞЪ}-oRiѕڂc KqVH ]};C/v(a:@Rk+ 6튮7S+Bq1٪f1Y5,;莭3fx8zlf1iJ1TShOu3~_>+Y?`$~}clQtr/٫~mv:[rACN~Gד+loӿix|Ekۺ|ld{oj]ϴˈ<{XS{g24kgzيV#QL]j54-͔@/Jtut3N5DW8f pp ɷC9[f2]}:]plh~[+B&fuutX{ոTo1~vqٛtċ##YOOeWZvƢjmI)t))8p-q棧Km{U1gy\#4:0웿<ft!ЏN@)A(.A:F+up}! C ۳4䓺:^ciZ"e>s +޹ٳG ӇԳVM}u~vUZ&~r=i2;f4~t"ϯ?[ԻG¬ ߮~Ϋ|v&783^lz[#J3~!p/~~r}::+/cڥ/ SLCg]I/}<3ߝ6x7R_\?yy{Ej7:ҏ.kn_f^[۠Yڰw|__j5"zy&n> *{\L}{=Vgױw-mI ,w#v>' q \lS$͡(SP4({XfzY& lHU5~6,'hJZ SzHN %ADK"\:97&?!\s }gI^F5e6XNM4;ϯVU[TAPS RKZ0 a:Ɂ0rEPhox "[7Z9'^:-!C Jk&DJ@bBAk/ .i1t֝2He嘘%&#q_e,$Ķ\#o 3)"Y(_Qp#D=<88dk'dM?ʀ' *U=Mq{Y=צ_}pܪ] /%G (lL~%Tlu>D6?Ak.8ñ]{:Y6: SQMt#JLϐ Z~5o'v ݕidX{Aq2g1؎ot1b6th};L>N_CχU[Y'ײ I+:ϧU68:!Yv9{'oTڬ 50]m9-~Ĝn>*p<lq*΋K;둎Eߵy 秹):/s91w;;vVq(clYwg>z`v{u 7c\lMPy8?+.ƵYiNBZ}]"/],_]u \YO y<he8ygLD^jQ=z&HʼnҦsH;`,=N|88!->k#Fq#佌[N{h|p*Fj# :g4JR5"Ѯ$:΁  5>NkgmӺ%w h썇n8ջE <J<0gA}^Շ[.٠χ[fgqd kAF_Ӌ釙 ކCॡ9&%\ohz{m}WDnRwbZ4Śn0&B2̚E5UCq;D1-jgufYn1k63oqrc8opًYLG6mo} H9N>upyD[!ilNeb켱@`SM%+ =#UmD%R#WcˁHj)Ew׺ W˰ꪚ̦J7xuJޮU:lz:Q#קlU%ǟKMt[*]i>7tgKX\P5 ^5oOdO}˷#c>QxͧTHj,sCHs9eC.wWe2[Mw֭_1C8MⶐM+꽁 b^-,᧸Aq{"R>WB tyaNjMlpgw$WRJh(5X_JkPF̛G:ftIB[!rсZJ&2b2D8)1\O9u,Hə*.v!BNBej$FcD[YwtJNT5]A{Q1ΆNv=f|m5e6Oi6]⷟n6.dErJ8Ch"jH9 [-%L J<ÅHxRݫ/_E5-RLYj)dJ ji+9kSs{96K}ISaв,[5 T:wJDEiBKG#-YE<M4Nq\:Zf$RDIbDT Z]+Įi[ILOT)JS\TIV 9t(V<ݕnXhlU8ֳDiB1괰%'KX L# WQA(ד&1^')R48ܙ:+HQy0yG 1둎>6dIK'klr10Hvb!)?Dz*IoϮglgZ:-ϒ䎲oM6\`U5hu%w3',8igvp%?bǩNv;nb8 v:7|O@!2Fɔ|P2_ZOlUs7:(OȔwkIbVe>! FbRFت m~⸹#}bC \{U1KE}:=ϧG%Xț6'RCA,b5Ytz G}F5>w֘xv᯵WŃכFgP2&spFg2 H1ǧ_#&wpBZҬ%.vt mkF[$7ȴ!'|0"&ru}pتnm.jXI0V܏K $Nf>W}Qb,6pŤf\T"Ml8n^~8=Cr߽7_77߿B'y9+A[p~>;&VbqJSػNj^÷êIٻqY³ ~-OWV(p` /hi1!{+dWMc{˦4I^!._+4_j黼QH;>7^ +7틽ѭ>m$x^H*Pc%A0V@v1pCD` F0h(4P(kٰ.e=Aog>ּ"#Z&*Ƕ &-=6$y`N'{:U_L%Uzg4ql;;_FRsXՇpp=K0ޓU۬T\KT"yy-MdI&4We/ BƋdtݦ邎1.-m$CdM'GJ0#S3\2/3$/ߔդYC2 <&Д=ڕܖLrOQHBuS`<SJl`^Q$BAB` qIr+8!:@FnЉq  *-)4ׄ;kT0IWM.s0*眐R|3m#GEC:cۙ{ fn0Rtt%$;YbwK_ڒ%ʖfIçbԅpgG#w]z5C&l${f`&^nnվ_u|DZx$f 8xb(H6*Ghn%$yR:~Sx&CO'!7h r dIDU(P7\$FQ/{)<0luck"G$Xϛ)*}2IpvRy|$$jp{/To5iq6MG :DJ[vC^@, BK%@<+p[~xMW0.mrL%@I{HkMEAHG 44{:oX? L7 ,NF̯;{JM5·:mͭ&[]Q&zziWMڤmښuV)6kwUf#TYSؠ*+'bHeE-3%YV]5X*/G`wSlSw'4`$矪+͝pA]̭} {:~&qٗL.%GJ4??4-l̮gs8Ԗ.ýG8CD)Rz}Xl<[S2Z4xn Y0V~8^Dhskk*2q3mk3&fVWM둻~E S}a]œNeC? Qbe_vm̙gǏOjύ8қ0G%SdN ѵS-qO]=:ik-^D'@j${ih[4Y-Vޒ pE:=ӧ54,6FgZ\Էt7Y`Ph*T3_EOFh}q e#m)K"Z.:DK YY6ɈTI1(&. Q}u>J}ȷ'W},y"dĢRQ2Q`@]|͵d$#b j3Zە OT7?XTupj-R -)az)>.{KUcWFz!RVg:+`HG+"U"[ymdY`zv, r_ MZw`m\2{2R9PBKk p8/"B8K{E锐ȡf%TIn<==eY6\?\ۯn|QQ`4ɳ@|i kgN]]?_Ǔo,9CLmvI`fSz>12K+*kͶmz/D1ro *jFcr3fyvں9UHT;^١bؠ\ڽ@hCۛ p&eRr.s!)c6%"Xr-[# P6p^fZs %Չ"Ѥ.yg+*YL &Cڮ`5+/˅oߗ)!2 %eM9'pWG)ȤeKbRTy $hPTQAmSl7KXG%ьKn@ydZ +j1q+*g)RSjnzP|m}ua- !^_L]P2\ϭ☜>L~8rź/y<ᅳ,%Qyp"F=!مR&Q d<494MḾ`[)=+2ȴI꘬gA4) !qG.!HFbGr\bUaRm̌;6ᛛ24{wYfݍh469b+sDJD9t$pA<#I38ꂵ:ܮlMÓfƞpg6A D"Bێ -J ڔNYJ'bb jڦG>6RCJZ`Lh #;Q ˩tM;!0%Q6AJ-ddT8hĚ$(1 /8$Q ڹxXLxX Ƕh #GG灋V*S$^*I Ҭ$L{&/mm %Di1HMJ#т3>zTJP`B6ÄNS[aD,&zD=l:pqb.nŷ6G깱(uYr>0 h 2Q4"Vĸqqx(xXltl0p&d>#׮mWJq 7p_XDqz*l8CNt\Mp>ɥ+TxM+wO,rбܯb !Q4.R™QXmIqE8Җp0 %_Dr@@q!PI2^ʭlSrn9S_lm!W) .ǖP<2S EhHbH5rbգz,AE)6BՎgy<7v"U]Zis&d;C KZqZXf% Q8(<7 )}^y(o޾hp3X-|Ij47pn[Vo^ʾxciGF0sC-@#gi=WY6 b=Gӣta.rzf\&l!^;O; , ݻ'8޿~ 8.>c~n2_iiJwH)E-"%ox-qk{1޵^ӥ<8yp{HCۢM͎)[fy.nLPcCv7u2\>wM'Ou ;hB3xր `׏͔v9u aoen.#Vh\O.KQ!<q| Xĵ~Lйl0̖_ VphY8c~i] evQBQͧw)qI(VG)'?zev=B̻$.OюT%v~{%M gUcTy)ץz=ߞRe'ۨ _%j)` swm$i7XR]-ds CvK~SϧzHZ() 1` {jf~^8 X|Z%܎* B..,դ(r]u%fg|\%NH'Q Tdx)]c(w_>n  .j8&\wIJ``o 'L \q9*Jw*RWo5|EH\չUԛ8χ"-OUX"9g"EJzpAvFpE#?"8"-j7 Wl0LfW嵋6Mq]et拢tOx:_uߕ=I6ԔNM (gtjp9"mns +'7PE< nG/.Fוj6FB\~w.ґsKWT4[a !9c&{s10~!1\NI+V?C \{o%;`ѸI70 YQ+bCOP6oa~z ̃V_*kUt֪~ՠ .ՇKIpI<:g-!]N`$@fMYT\ԾJRJf!# 5Ƅ,WhLډoy<> )›ƛi{j,y_p U<ٿR'!uRی/lUR>pBQV@G3l(CS|cnz5T/;Q@=oz1{1{YɜqܖI eh=p Q88y(   !Vۢ)+9/ͷR͇n55sIZ4bX:Z9ޔi$*Ȧb((]s㒏 2907͞>J#뽙;I>ףv_[)'g.\k ڭW#5Ҕ^3eaޕ-GrK=-3 )цP\ "2"BBʐ䲌JKg@$gd&2 8L8nh]hn?RM$61V }? s)j] yU:jWH I IS)/xV}G\E@V \ApD(ܿ0``s|Fp 3e-c+<2w־Bb 1l*3!Rtat2-+ &h:*=kcR&rIzܜa,9?Ə^=t\v\Kw?gFy>7IIvmfio|q]XۂXc"ػ[m\4NV,ܲ_F?iO6td>m$ڠy}tժ1C 竒w s%p$; Yܺp6!XU8kz_F]bR'<ӧR;=Dv+tWeԎw*y̳CACvizװyd>xZݩ.a}rִ}j#ևBe53-)kF OKk+Ħ-U=m" 뀴){ +{.'7^ '!fHh}7!O_ @~kWT04w+$ܣ=y8Fpqgt>,t\X$-]=]f/1w9ݝnv>F')LǍO8y@W/U9Ήz23.Vjbŵ{.qlx]fr5V6cWn\xg={BfX%h$blxYE"ĵ@.e/PSe!}ѧ!dۋmVa c(g .J0#yPb"p:/$WJ^EZ~}&uO]ꮦ;KdnnM]W䅽Ǫ>G{RIz=iȔ~Li"ɖTzYij4Z!FߘB4080lhog(ڧ*Iᒳ>gO`=};I&hJ1oMgA k()(3F؅&`"C)r v'[/drvYU#ѩ4,)0Yd!3V" = (^-uź&Ș yoh:^[3~໾J9!9سRE;,vUl1c93d Yb6IφW l9U|Wx8W6Y(8:sT>)=wbSQ]-x V֤ UNiD0,O1ML(h2JZḵ=\n% q|r2ӇIlQާdVf8t%WO4'^vM>ݘWz7>M;9evjӒ_3_nG:9Pp9 ҵ~_|WcxsNpm/*RGmsfճLp6r&+P;m]R&dne fF<H{v2ٟPѣd9EooY |2=:1eT;asf bVj5*`"EuJ.%0xmJї18-H\dՙ.{gʣ+VPѽݧQ?Gӿo|F挩!$K$u)KIeJu9zLy𩪡&` fs$ 6ЕϊO19xHHJ$^kfBeC&㆚I͜hi*n^֙o?%ۨ+6cwϷqluyH'mO;|,K;Rwبv%P**З"HXH[d1πgA!hnb)t(lUdDs{VV Th1[/K NH' \&yqjXXM3B3`Yaʽ-?fܱý:.nuuɢNhd24/8b2&d"M0t2"A22#!I`UElidz sQ0齐DˏͪLs *+!Mdh0>@*RɞU6 Dx]v,ݾDwTr2YփU'n9j&K(.!d V4Z"`M$O[,)S\{(oy_M C{D'3*oUeI"844͊5a{8)!eC 2ʐ'MBBlD&R3 ~2q4^ gAF NΘR"KJH O֧LmΞY""9"'du [yKHDDIWmlIᣱ&A ?omTDɨh:s +vEy/eZG֔ެ}ɟJPVwVūpw/n;Rr쭣'4?Hٴ2\ZVgj?ՙJٷ[V{gRy'&b;/T5DϠ7'BiT&t !JT;oN:,*-WcQed,t0VSW&4Zr2 Вu5zNLߋAZ2v@26{)dF ٕfmw[Aۋ'z{a7qEarv0`"\c$s%+^ 茶0#(qLJܽh*€ԘŘ4JhEfLP,^RT#QA\5Fwi[m .JV3*3TEdE 5;Qښ`m#A.ypP@TnCHh1A}2]wO>4qI/K8c#2ȸ@@&@ @F<_Q+er<1/::hBUj|$H^JmpZPcC֢K>brY.xYe^Rf.#9/#@r\n[xu{\A]Ł;tYw=/`V~u^ m)9Ϊ2̧}} dk8Jq:ܗ켘*-`ⷍֻɰ}⭵]:Aݯ,6C @d,#/Px>ӸPqub㲪B2qS{[=:z?_Z,L_[6Soeݱ^*oxcm/4#muVc8twli-klێ"5E{뮲-'F='ȌM"M.]Tۨ3č/V Nbie09 '{0j^QZNfUoV4nYoGݲ,^ բJ{jI%Ϥ⚉LNj>؅`)3( J\Jlꍥɏ'աǡ#t0q{,3E?O:!&yS1R ܹ.x-m8Aeo^Pcp[*PHV'S.ygCO>쌜=ʭU^^\r$QEH+ZxZϨCR̃ZGLR;旂tϗȮ;p2g_2(z++)z_PdHi鐑(09Y9gCt~y6qNEe .pLK=6%}VBC :DhИ)+\Gʆ|P OMhl}PiTߦ]V|PQs7!6"$d-6 #}lF|eL2 ؎l/,%è\3Z6hu^\h,C|vcY,:6<֖ͧj=WP'%GYgfkɾ 2t.TO;'y֛p7-#3$VlAU[Y{N&[ڀKVҦu"U^l:ޠHckG%?.ۧMwLIݬ&đ%,ܺ۶̫'_[10\ݞ}}xVz׊G]oI4\fs}7OE|Nd\|w#Cpq>D6G\8GSwqv o}8ş?E%'TY, u 80JH\;:}@u "qXh58)=J59FY[CF j$k`]D`G7A0-d85>KdZzK24!UT=%Z8##\KA'"J+c>KQ)+3&'I֌ɠCJtT3rvKF; 쾧g+_sSԮd_{OYjl/j#< %9EiRJ\D-_tx@Nýh*L(1ǨENddiaqJ傱\H: APiĝ;-ƌnJPSh/IK}AW`>*[-ApxH$vJ]6q ֒ ajWep"t5:EgJuTy&p#*u/t.Ĥ"Ǩ7 d>uŕ*N=8 M:J{ѷ }+QYt>LRhC7I)$3Y Δj<^t,HIxp$GlWTe[y.S΁FX43K_DZ&ވ;R'Ќw}3N/]8) 2R1ϱt1nz5r1Oh89y-(|"SV D#E֬XG?8[d93]5 aPN%9:ӊsÑN3</`hŵ_dUZrB; Vy68:aKr5 .f,k"]Uѩp]\FTez֓卾%ޱAEs_FwnM#4b[s>`r217i1Z~.^7'U+p4{joׂ#hj^#-7w o]I+]-ږQe֙- mѸ:r""3z<^7=]9oW52}&VmWU9%?ƓO":w'zˤ yz\98:o}uys껷Wq;0KhH$?6VrsayW[W`ZK6lU,`[l?_ѪqfinK]/`uEܲ?n+Xp#İwtyՋFxb =OSvUg#ȬSP0F$!L ZEkUuTxF7 }lh?)`nb*ȯ7=}SMtɨ d_! )Ij;NLa6<\xp}r7;_ꦗnU}po(;͛yZgdX3x3$'2]^T|Qg!/_qӷ[1~AS&p )Z#Wϸ?}èqmPFz:\T)‘J!ZYozUˋ:7B}x zpI6Ib4:Sor[ixmzcu64:)~?RE=O`,c؋E?TDP'FFgvE"n<'x@LWT{VHAH[x;7fɪcRt\5VENBbc)" !_(uT~20HĖ^gz_1idH. 6/䁟c#+3,oQGe=w&8RnV",2p(wG/{ۗ1}ShG;=N߽5}i0H'SS-Vh~;AN;;%wRw߸eSg΄:O;MC7|w4ρ3TY>ÀdoL0[L4*-.{+ʖeۑ!E5YUp(thq}+ F1/-h54 bZ G]UZ *Z#NW%鑮 ])W V+LCWHEtUQ8銤Rv0tU*3hcWekbyCWV"!aZEC+Fk^zxkc[6;ۡ+'!+f8`kCWfg!=OE] WկDWC=+|]HW6=(-A *`mCW.ڡUEKXGz;t%A1$uUK7p ]U(NWKvHW说:*]W"-eXmhY̮'1ϥZ񵞋/!/?/59bmp%B!4Efh%RuQv\&|!7-hRޙT)%E5EN^UWLIuZh\1s+eI `p+Qߥ0G)`YI`… 4VV#]AB@DW]1um!UEz*J4#]A2U`dTyꑮ ]B I]Uf8Ir.P誢tUQQ]E(, *`U,rҌ-ҕCJ]1\b(tUJw(qvT ճ[xDWzJ~ѕy]mz@]UCW!^:x( ҕDty 4Ǽ).&ztKSt&iBP(B$ʑl Z+ɼ;QhSV왇|0JeAK6M)F 4+Y>HebJr紕\"RIbE/"@9k`;˧23'2ٖ2jOViH;.WO#FT=_}^h3ΘjۮHǫe 9]QΊ+'2*'ΖHd&dPi5b!d~tNw][]}iz;}.6AGd_n3nSږw[WV]*"8F2dzLjnm)tEO7pek zK8/B $ gz{7 -&b.^g'pgN9DY'Ǡ>G'vrm 9Y lL™_mVD3ͧS|>I6o:;d@o1;&Ýqro=Dr'6|X\-9NMՖww;0qb#<]T&t6G1y$c0I%K:eRX v궈Β*i@%ϞgWV-?\YQhynodd]>ϭi#B>LkoN>竓W' S:r:=96%{2mU駫v$i(8v鬾w?Ο<{Q[;z#tx9GW|?V?Od~Xɺ}&7WWTGiH8 ٟr1XKW2%abfϵv:2J9 ,SРB>ZgE M%x`(Hhgp]=Jٽ48AۑʰsV%kV`)Hs {#笯͑UD &Ϳ ZfR(J3BQT E!wY rR8V$$*k5;jgpva4:C0>| ,~"_.'G1 ~}fŨ׻\G[ًO/bE;!Wi$=W~V8PRCҎEJecQŒ Ĭ"DK֩xod 06;S Hɻט7ɨҵ z}Z ~V8l4".Y5VrT~5ӛY;nr1X ð>*wdNqs}&S3D ,@"JAar~+&=x_zB\cD1ZiȹId$#u`_VgR+.&c4I!bpK(XJ"IZ,)ya4%r$cHI q7in]ā!T[(:^?馳uޫL,ڢ%JQCDR71>esJFonDrS͚Loη{ f7aT2(R P| A(lrJ$2Op{t1֫w:CʾAЛyXNC!m2K!Q3NndWJ61 el`QVRbe*XTuAbEK~AG*baQ:HS1k]mذxM09,VcF( 0%)k>.$JJ'aQVBB^ٚB#tH x(*dҒdM*E5aJYp~`2TCv@BSKuuDnȯw qӧʂXQN}XO]H҈e$K7BKf;r#n܁a|v#7`"\c$s%W4T)茶4#o,qLK܃pRcrycr(R#*YOQ D"s֤]ݹ6ڴGcm .JVO%g. UJeW($<ىm t?Ā$(σ#BE dbBeiҙM4&ȠOJ֝Mһ&SX i>) yLԀwDF(lDȈaL]Kc} ɃB޻JZ$E1f:a-~'<$: >DB.F7bО=/ wy_na3d:L{ǜ ,8p'C\A𱇏%_e]WBDJι/d%#()&jo@ ғc%-Qy0=-.)3G/+7[5rolzX^'R!6(׫Pܓz| "E..b=9c*>°b|T~!Bb۸ȕ>}l58} m3絶T7hZ+0g.Nו}3dTWE;]\艹vM[t87-#'S8zesRl)-!Kڐn;ɵSbUTyYG6ߐNKl]n]֭w9ltfsYu-KlYvݮ[;=/o|E7ZhMw{fϣB.on>ʚsdzM|k646qu ۢSR\Ӿ"|7ʹCvgI*)BݙZp|oߺE%'TY, u80J{O\:Guz/f{n&mZ!rzZ~W㓮T<KN&$SQjD gtk $R$Zicκ\8yU`)U: f85c2(PR ]l:%x% 8-n~Wo9W)nW3Y#p, ?׎䠂MJd*䢥&: =Q?RsZTJIfP.% ZOhM{gٔny,K=K@gʈh5yt,H 8xp$Gl&WTQ/F˔c.<& R9%I55]W'.ikggʘg ο59lu%\LVraS7$?B1T '?@Quz_bD|(ne$Ԡ!2QB9aD+)n2sv}0^|oL=ZuZ_+OwՅ7oo;Ahb[/a:kbmo8Gh{og0|-i֑\=y0mEhfyBa8SĞEFgBOcSJ+GedMv+jNsK>Cr(YmZjT5*; E\vo~7߾?߷߽OS9m R^$Y%߷6&?OEW_#N9BERp4D0hU;Q$T=mv64ʥ}r8e.}%x7.1\N>J BF%JBQP+wNw:^^ og{bx}7;_畽aU2ӝuϓg>g{%+mіݛa>f:ldl]h\(GuhR<&<2bV:.yR𠢷>^I^v{Fنrc=.qBso .yNJwR ifL҉`UɌpėyV?=^ތyL^#ONI;+9Jkj]$&70ϵQuJ{ ÅH$KI.p .6 R~I,d`$PO1&$ƿL+Mxbl:Qm-OI|zKkoμ؂n aq-FDՇ`}Hb$>G<'a@g'SPh*REK4ErPuv0aLØ#1 c'HQk=Ve~┏##J1gBCR ^0OuA% peT7wROT):(=mk7]CgX}) k%u3:R&ĐHpҁ%UZ_3Nll>{og#^t h |FQ0i!2ZWh83;&Ō D(Dm6⿧4T>k0|˙5M_1ͨ}eӗ{i uZf9^"$U9%B$,cEtv^$+^1ճIieplOL}wկV?GA oϏ׈ru\C:4O]Zv pĐA܈D.Vvr6°˯dDzHo wߟN'w.O,[o-x[~=>0Apv_%O%I}qս߇vo}LK@ZWMu>Y%^KNqԲyӓv7H3KPsu~q1+mS !_/LG_/ۊfSE`"4%\iUlCX)ު0ƠR%'Bwg=Ȯ7_3ւLl<߼鶝_2={k4WeWyvuo/Xmd#oo{hX;G:?ͿŲ j(&CtU;Ց /C1 )}DI:dOx=`HQןسz8j%$R0lmEdVl=(t{&Sg }O{E>}8ͣY|]9;(6!ǣߖ<>ݢW~7}dzFs&90t0t~o 5]uQ8vq^ :̧ ZEmөr5I乍ʥ*h_+X\]E{kuJ}کPW3pdK*s 1zemQBΞGS75vv#gώ{U͠]u@ͤ\6tb6Cpiq\߄QX j@^ǨP$$wf#jr :Tl!!maBˑ8un2B▖.[\ Rɡ)"etq~2J6X7_“WcAW<9\r'߲ڲė_JmWxb"@k /Uy* LjhUJ87 wՌTs1Bx#Х&Gs(kJ3A)gC%[3v#gftӅ8P]ȋ.|Qۮ>4xb{}rLnA]}Z__0'\ccNՅkI!iJTsVF9eDdv]5aS<셬Ĩnj %m&řW-D[5 Fns.ZwEk> Z ubD@@މ L太 -Ņjwq%Jը{E Lb6ѡSΥTwH](,cbhyVHVbQ"BVצgb" *2SזL eS+pG":^\m)ٍKҋH^|-z#N&h{F5_) 1ĤUXP;*Z+U)6~ыqǡPuև0}*lDFm. ([@{սq4%b.ޞ':M ;;ĸ -:xJ).ڸ|6XGl@4v{ہ9uDznKY7ɬ޿ǂ{̺䯲%_ d#ث)2@FVslT^!_ $3^7aͭ% 8C1'.34hl6e:a<5i9(xgm;ZS%'\!L I[U֙Zq rp?!YJ WFbX86e!S땢l_]kԍ=D(5+wü}9BDs#$ bSvYWшLe&O3TNT7x>{JZ1R IlBT. . Tun8# E9;n=SYЄMdDJ-jFRștP-[0 u5F%\T㬑@WP *r)Ĉtn)g\cFO!K{\lp C4$yDn'&L,BI^5tYDFQDF-/X3kbћDZ$"ذ@deqZt8Zx9pey@\CM$5sʷAI"bΰ(.cU>\̦*Oؽe:#z L!1UcɐUm] .))q]Y;EeDehgҍE&|{Gʪ T< 4Y\hr$PB8UQq2/E<ǵ36+nrttZ0VͰp:" %ö4z#  Kmyqh <9r/GOsojCk`-AA"Z'De*NeZm2F9m jJ{f/~A̼lv/csۓ3{TPBV˧MN6U]T`{GLxLw 4;}{!MĽ'm s56i虼# U6 OTKD ,9 MY#*+WIGg* v vR;ꑍQE@d9zqcЫ@5dŠ8"|bH.TcRm@sDFn#VWXw )7r`ت)ul9`* v*!)E%Țڨvrb*PeE!zSEr52ưt{'o9BO&mbo,9A[X,DNEDsə"c3˒4=jĐ]۠SVK ޓ Zҁϔ4h>N:p;\O`0Ԙt\`QuLҷߍGEv>(q*Ҁ 9]9G(kVVV{9@>r%@ƒ"kmH66'ڲЍVuhQ"0 ;PXTbs}ҵ|.BhRPeD@lcU5.pΜKc#I+w;R=㬃AB?%Is(!UσCRP1`45_Uw}dYw@za.e2x >bA[O(Nbx܃~iΣ>Kc} , I֚r*CvF2 ˘HְEac \ f9yyn2^N :}͐r3cs`:(?0#+ $f7RېDh(|F ʥ3>(V \jA-W_Q H}}C ӎ́\`[޶DE,m+w7i=S H's,M'.F33|%gS4a],/S/eu0Vj+%g= ?_jJ{HLJh琪J{\.U*hOB +A(F{qqqh=qK#.jvRp vt06"X> 6:tD![E Ï(:o;cƌL"^ˈiDk45D"4*7FyYFStt rk傪Ֆ`% QқBau񎇂 :Y;]2(fJa< `IՊ2Z>+٠RzEĄv6tP opV^K^MΧƇ3Qxp)pBMsM-:F)mڄ~=~>ַd"fx.ǣl~1) 0|P+ߺʗRBSJ!%'pafU;nͳ *퀿A U $Ӧe'kHQ>K<;d2S'K8 lYe,U{p96]>\,_kTùץ[?%|[?ZnߌД&7o;]^_*M M1 Il*-4EscKuJv];kEJF]^'&5N~TFb飠tyUzs0bK2Y7fY!̡d^Kn;5/ oN5CK%7t<)c"{:ΎLu'vͮU|nfU%wliI4yTA{cXR= y\on쥷u/; 3XeгJf \gFҝJcIekmWR mELٺZ]Vz$ϧͩ5<[ܞ |\'G_ɏëiyQlR>̮9?|U t8:Mg˷. :tm 9 ۫굔 BM:ї$uT!g6ιUސaᯧ/;|2zփ/S<15%VĨ 'H21aҔCIl,3&faFzڴ'y̝MJײRM^$O d`t#˂ј@u;pjlPP-y`E.T6D"5H3;/ۄ:R")9IEdJafXTqL<C tv :wXMy LPN@dG#3G"Qa4ZNl L% XPlIKۋOxhd OiTyeI qfu6@&ג:pG:^Hи3Etʋ Չ[^c XR%Qr_ǖHa$ϕ5c擬_NYkg.Ә'W ϿU`]k M'Kr9~Œ<ɏ!M0u$t٫Wiv+5aGX!Lj`)"Gtt.#So(ˣ0yW V" 9lB\:$UX]Bw YT,IPR!c ҙ~1dS*D^7-'i=99]NV0b拤<1JAMwX|h >lʘYU<Ee`v3Xť2[sFW24-|ra!֒-1˶fH[3!̲|Eh\7RrGÛfE}pЪjmͭ6lkm ,24k$5A|O]8fIm#_G[)* *ǙL.@o޿K??9w??{&= 8"ۺ٣X3|*J3gʸZJOﯾoEEGGwpOV8-|?e)hi3!oDb;4P`vYCnirۃ Q4@s=Ҏs wV'.}h+DumHmAE^9C@Rx&`m0L f xM6S;咗Cr8UC$ XeW&!L3sp{Q<ca 2ןtrSPh;C!ICOvŠWaU˰[rk;0Xdbjei7ȝOtpq+И&mH}dẃ[4%|{R/r^to1vFmD,DՂ/D6݆u!^+k)Lk0X&HSE57V"FMsFV;|r^pq>p ڿKc3vh{f #g`?`ν1̒¼ʆ#~с%7ih 5(O {e7c-` s))==ym!AHHhSxfSnT$N,'0LQ"JeycƿqTQL\Qj10rbrX+/B"IκB@h6(ꀢoXWnࣇ1TRIrՕhM".M2eጹ Hq2לp=C>o"xas=Fă1*"qE&ђZ3}DqUT qc%8r#zF0"7dQ=鿮QJTHay96Ö@^Rޭw)Mw?jBЄ b4TA໴!6RO*%;%OJzk[槥^mVkZ%۩6Io$5ͭoP,%3O&uf["ȴ$Hպ(]>oA{dީ#KsJ|yf-n!k2b.}.e`Xu_76Jo-uHؘ9]yAT.@$H qUeɅݝ陝~э"Ayq9חST/@t # L$/U.((XNHdo/eDpjYgDAzWDE)bgcOB߇4 ;/\{L G)!߮J YhSU!qrl'RJ}ԋydDҫ*WXWQW +{Ŀ.@e2z ˸ '-+uy0LDbVB3/1fLZ@_`R##P.K`|\ܦ:?xOS`W hNBP -u'@}͔2a>Xjicu=ZCL 你{Q}Yk~~K>蝹[mŻ6;rt`tVg5y7m d}goP7g9v*Vt/el᧜mt5pèG%сT EM|.zT 5:z _3vv޸gZ l#1dtš{/^SɫmyL-k+q ^M~Sh`We|fq5{ ˏF_͚rT4i|^V l3u_=1hWdU&WSWZ.Jb{qr\zW8wk$+K4.ɫIx՘5~L'qk/04gG `?;eX !9j1G[2ڜ@Z% .̿L|!xϫèNo?,Okܷ+e`~?~9ɕIP1\ŭ!U9cSsuț@Dff1%/wJ4 u*Ԋh(d8!0ɘ\CO$=g@v$T $|A&^,kQk6Q|Z.6+D-ռ0ZyŹ` g}j0.[_N#\sB_Rr{.:!1 Vdt&WSH3bin2;qTUV𮋫L/Q\ c;%'8Ɉ\s**Sk:#rUW HyB >zX~**S{*Szqŕ&8!q+z:IrIt=\\!h/^"Fk}:^Xzu j[>y>x&,s q9 wvx,o^ts b2b顋eZ\ ʋ~F,?4? ?%.AU>eg 7g$kސo[s.`{e<^6zPioUq#Wgۅşf=yV0D*cmq_>s!~T{!Պ󏮫M~>۬np!=k:7~gMU) 8mZ|YgakC ncovo>ۇ/E;G{/srmΥ|< 9%,FVTkVq r Nzps av{3/]XQׁy l]ҽ o ;\ZzKXR .[ S^LS ~!(@cܤJ NLQBO%T<0_"I qF .$Mgl>L//u|zN kA&)vM2 1mԈ񉼮Zfю /$;ex?\d1a In+@!32.s4x{!vrs ASmC6;G9mnvݰh߶?[Ul}n4䶓mXjv;y-t<_nwxJ\5$3zŴ."|3z5Pk恄1@9q}Fx@Q3lSx5l@y5]?( TMɆF} 1٫bj>V#koH2j Q_ur?^?ד֤Zɻ&L^;^-%R{]/)yq2zr\jeC$TWb~d:P}e=1MyV<ѧ֬;pxddTѤ =^=0vcv,1z#/LɶC<씍El3%B}~|sK-.+>->ARcI2@#x4>qKK"@F$%=%}(yR?9חY(@ dl>_zxe_/(+@(HS& +S O(kxE"Z"f$[s;Qʡ|6c@Ԇ9sd"RE]HMkLHYH)ʨ4@ G+S&N Q)LzO@*t1#JX Y1rv䳚~ l.yA͐%OK;^ZDc5|Q0!R*g,k9AN?$7$5Jy֔; F-D "GDWzSOv;^!<Ł\GjGfiZ,HNLѓޅ("8:E_LE.E&.j:jc- Nĵ|r.%~5$E)Q2'hj*wzn+Aڮf+]abGk̷{'"ZdҐhJgQ^YʥCA1F%OFypT\*0 -YT4OCH:6`"KsQq1A,xGTԺ6튑CKD S\">e}R]0jYuRwZ}OW8Ip 8UQ[ITE#BoGZE̢'j˜FI+'V ^EF;Bh-R2q+`o%E^kBӕskaY$AqWTDbI C.N%HǴD>)Lg"Zo_CӮvt/ZgU0D?a`q/x[T唄Yt uVZu,OFEX;yú1Ye(3F*gjo0I rkKq0r6p\>!~ՉYn]u%[u5⺿\žF4NW s1e%uAe=)P@f#3+J\|3mXu$gFi:ww\sh 7p{?~!G>Х=Jx=3&:lMŰS S\j".M6gh $/ظw|*Bv1qԋݨR_$+!Rhe$@QJ;h`I,iaz?,.QP#$I!KLJׄ M4[Mtteh=9z.EDoi\0۵|@CPpP$3Ú>6|ɿ@nztzh_bj} 'qQ?JdFKWQ^WHd"c%Q%BYho'5z0ayvF8i .7]t3l߽O-q OR,t,yκM,KfGtdBtYt9 :dɌ<$*J+G1~&^kyˣf).xP-a=KyG0 ȱlNe\AP^U~ج-+<(jb4FdU*kN$47ƋO{"?=?ѴF*ҡ˜P 9 Qd5) rPY+9 Yޟ#orQ9(QX,:~u,QgҶԐ ֋d [z"g3rb~Sʫ鿎F,\VS <\k~kJ}1ŞZI0Grt'o fI^?r.ƧY?0:&zBϫz^7@+~,R up.^{ tN/ gtգ,Fr2f!(EU&JҬK^q>nq&+dQgjF$E^w1 ]c+Dk{m_+5Y0o':M4ƍ,9BϯQwT%u5ɾQ*a=Ejԁ-ϟ\hZFUl9vNqѩc{x= X_>;w?8~c>?~8XP%BHIisnYX{+Z5g¸]Z_xlߌ繦\%j<ѩ屴+L?3Yg)GyEﲄX/Ͽx7ȑ'PqًX& Dp_ȑvܑgEw#Oe7)$@b\ї!I,u|.(}HdI ȓE <Yx͛!G# 3JQXbgeL5&*9tPlPt^OOr&F)Ak=5%)f""CqX@TҥZ;GK1 ;} A;ewv 5@T lIȗ@!<$k!XMOA`Rm>~h+\f*(1WTR H$B <BvǤr}w=st>C5Dm|oDed>XsR3=C{lFE}Uodm/d4ϧmEWa"{%Zp6L^3B-Qs拉(QNBbic8Zh[FSIT"8X3@AhK$Twp` 6td.=4\Rˮ޹,q$kUU̺(b6㙉pYV'J S$,{~@$EFBUYVWY{XEb|\,6(U?S+*Y[-'5Wc k.PWW$=bE=<׆RsNrUEq=},ɓ26ٯƺMK&}yp1ٷ aݷ=#s_ic]MJ=JoO0Geݱcچ/m5%lvcW6Ob}2 kbh G}|j<>7{1xtndezw:~W`=e?QO;3~ ]tNɢEQ!ؔN2rdcNjkbFCդ܃ >>h]*z'sm{vc9p\1>8J_T@augo͛_ON?(a]nwԻozZo o<w>X`er#66#HvٳVWRZ_Aim9"*4]/a;VψyꦙSLS $nPb]߆EZBDVzPV=|Q~<>05JER}%I4]Iu@ӿY@<`-xTcOU%gꪯy'l}V}Y)_8 I/qі4o*m<և[G /Ӻ|yxy1=CΖ\ҾԔshᢍ6){E6<8][\ͱP#Ɂ{Зx1}R|FkeB8n|0j>n|\cֺ߰}7p(^N7>` W\%\ k#{pZ W/1\1y}>\of}uNj>Z7|Mnӷ@m$ڎmӓ'R-26/MYg^~ƅOF_wnWJ קI5ף3k5[Iot!槿Q3L糦z7 4SnB&"ǝY)wYv?>ݫJiہǝ|]_7$@3l_+Vo[+fN OZO(Lzk~<2OnT0k:;5#{]+e/io=p5 `%ŇF{VZ W/1\Q+:p5Pհ}V^LrP0`0黊 a.C WZo=\ +yaI] ] sLr* V-ۗnHr޻?zl~x\1"cXMܻ (snЙ%b~Co~:^ԝ>\ߜwuvG1эn L;\øgPISP. TY*G_6o{ QB,gi`}l}vu^mf~#^]T6\ܑS@w:QŪ@4w?aXfV~:Ӽ6,bnwF<4f|`2~AKvc$~:cKB%Y 7킭u5Y <7qw!iWtGl[\nM2YoR签K1x^:Im_KətQ*jzfrLlM:.N-iFﴁ}49eҡYUcCSVS ڔj͠ZԽ.~k̝sX\$ku`?5n]|!hZd{8'dRWYGkN1R <rбYLFDM\Fk۫N=hS|6nZ m1Z\G.8ޙK:YiwfEa A裃"g+*tk<;N}z*^13d̉]șbqO |jYXUE{r'gj*uϬǣ $]1۪TwAGQ0QgcX2ctSt-j I:5K)%5o|0!4gmDeG^%¾X']4/[ 1dTW)鞺5 ;d1KuXXtd 5SI׆R6*yˌBi휱EB\X$S1ٽؠt@<)h-a]K68E @RACWeWP bȳեcMpmlt%F0Иa'+3\5m *Ae8 oTΎsGح+߳e]w!t4MQ +in%4Z6` ؕLF$W{HQUF2|x蘊'8$;e5%t_8G(Z RI"iyM*C-l58tqXei(^BB$>y AXsУ%ԭ-cx $nGfp6EUUS Yߨ"|U# d-PBpߩycM0~HvM_A[yhcG%D|uj /sJg۹DSQAc,hX ]m!򢚑croվgbfn~Iubl]NX-ZB L{ctёpiЀʬ!ug6T+u[U q/GTYc!-#!-Z6Hªl԰AE}+G Ը1ĵU FsryXy=znzu ~6o^FextlC!ݸiD3 --z ס(?JfY:53%8O ]j31 8a/g#·^TfĞL:=nX(!/[trM1WdsC<܈XQD{D+1-TP=P댂0dJҌgOt0XFշl0h #*TOϛ[W TN1֮ ur3pr)M f bp cFYTPcDi)7bQwXj"VNX`PrAv#ZqJPdCJ,tRՒZ`4b/(繍*?u+jp!GRcQO'Q|zzs0s6!iܨƳ2BK+m߼|: ICLZa6Ӕ;DIMi6CiOfO>6>E?i8 ̇st0IZg +C%r@H8 ငH8 ငH8 ငH8 ငH8 ငH8 ငH8 ငH8 ငH8 ငH8 ငH8 ငH8A8hNKkb~8|BPD-`V-\[=+l%`K>vӷc͠qtrt,_ 7Hwo^|uvZ/vDo'͵r!bwmYgviC1Qf&#N0@bASj"6i[Y%)jIlKuz{]ZFb3FQ49 'zRܧt=.]ҹ>b,w5NRuo_y0u"L "1ȴqX 0'-zzMҩ&GC_ݮ6uhڅ`訬&/~N K =FA>{A uX RwfN uO D+BWXn%K8Sjm_ŎK[8t7*Z*L k*Y~DT,pb2<͔ "cpፒpZ/=rozR7z辛lXxg== 4457|IX?֏↷sj6;!l- /z^h}/+Zs\AK/IvJdeW譤aXǨ0#: I` "8e$E'KE^h;(Һ*4kKn[~TT O:A6Og!;U?}j{҃w a.y .f3Uz{&3OVܽHj҆ӫEgu6,xp*!jlnXx D p$ cfn6̆Z u\,f9 YYR 'Tc@<4t16Oar%6m ]w2Xc,1*/J;ofpP/$r[kݪGq/FK%ZD:Ǵ/Ǟ>|% LAKeǦ: |h垎@—DU4ñ؁;l ]s:B=om34Ŋ_ۑ{זiˮ$.Loj~m>9GEUq*[WbS?>f@aa}񚴡KRGQJmƌ6>f)ڎVi6rluVҎ 508ej;Rc+ S&ޯ iY3G{+؝Z}fƵBmuwm67&40r XhRo.OayCl6ʀ>_ao QNJp \rAߴ<8}6 ?6d--Gwz UIcMc;ÒI }"I+Sћ`A\5Jm\ޯ%;)$| fZ tka[ߠ)AW KH2w#ӑPxPpM:Nǀlu^]{jͦkN+a:6Go}X5-TùQ~Dp:.ai|ݾ\d3*eP2YMWmzoռ,|wC5NhK}ۭcޅ!fԷdէwT\tYՎBeQo14\|4ndeSO:+]c^|I4mϣiH`״w P]k'~ކ&΋#gS?(e"#)9[M͝&=&n{$нҾ&# ثP#"<1Fʩ!tĆ8ДkmC6k;;:L,pdvځ?Nv9 f7:w>3eUf?%?M.BwiI4NdEΦN$tT!c N.q!Vݫt0:{x  )NĨ 'Ȕ7ƄqJS%JϘĚ9ۋ]ְltoX {EM!$YdE ":'N[*%ȅʆHPs lsϝyqrQ"R"IE8Rڝ#-o}Lk94G  X:Qbk@yL6 K$Z.(xl0%qfڤIA4_f*zW>Tz=6Hh _FUB>aIfNz?}wiVczҴrI "UZ{v|n%OJ@cX 2`)"ct|aMQ+;>.@gJ2a քzK K(akܝzy0Vb| }.S76%)Kj[]_ɅNNӫqQ#V)T khe/6 #LΗ0LjYomLu^Uxd kk BJGG w ;ː2FIZN`E$SV/(瘀o&xS( |Fw9lVJHH!eֺs7JɅ)F垷TC:bKVa=؝@Fn>;LG>â_.H$/o-@p348γ P $AGJ8 <?8 zq@ pCTkmCtÄkь^[4b(#eb١vP plC%)*ےF҅S-w9.byvMuf >:&;BL}#}\B6 Ȼ4UˌQ3ϐ,|9\06&cĪQ+&fĐQj`m鎧GKɢ2? ԘU|P_ߨORjӆW1?3Fky1?-ab436 MKyo:R!?Lb.N.4|GRQwX |_&3՝WB7>xŕvW\.## _C:) /Un1 4^nc淅 YSE./VXL_. XK g-7]v"{)6WmI~]A_jUV:0YJ@Lbr޺NV sSg}f eg$BGuѭVou[.M0?/d: bgȻ},r36_y&.O_*́VٻNw\9/p1;BՉ/lR:?,:kw btN~|Mܼg7Zxri;޳-3}kVon=r㣇IsRuT(E& X|DUg3"4BJ8λ4>)O?|yB,j,<' Z"HWɘB@lQ.^Bh&=0nPƞ[Z>guC^8x % }w>ihrѯ?z5]Y$21u[FZ Mq2v9vڡeE%g?o[gO[!."@X3V2m!Z(7.1›R;_bb?a|ڦAVzv~)}v} ?ɯ/gA*V{W)MUF ^}\ȧk%"GQCsn{gU}϶E 751p]~RYQ>ٔGѝZ\L"$V*CGh-QQŻ~+wٷ.M/ %Y?76ݳqٷnPd|[>胃xz;-W뮽eQlʏ(S ]TR׆z}tU &E 69"%+\֤QԶq\c^3Α]BM̡dW+[2 Bh @5263g72*ͰY ݈W-.oZW-ֻٷ_vmg/yBΦ_NfeHA_, * A*CY`!EM۵t"BbJ2l*ch0[LQlWIF75 ɺٍqz7B̡v1v#j Oq)6jLuU+kb0t]UmEm TmlHqmĨ (:b krdd."QE bChٍ}Pe` "6"o~D`KQ4R8o  Q-NUvwHQPC0)6.Xb ee9SSf$Z9IuU!u6%{UFɚ .K*Έl ϶1&$EXPj[ * eT/Cfc_<P쇇/a;JM.gcX?R24>gIpc%k[n%4\ΉmB1AcLL5 4t^oM3շX*4=?w˽e%)E.C#]11'd`)ԗܮX忛oKe{,Ѫ~}jKxyAJEUrօb=VΕeZkr9UU\.U?g-()^ 65W"U!lY%O|I Qi|9dA64u4vu卟ҫ^Rrܢ'[a L.K?^!.-:! I$F C:RcA 4:%]B*ބQY\H@#r ѢW?G;"cӬ.+'A/7zV>ʳ!P0 FNpm(MfDYa[t 4{pXsc9""qlkҀ1;X,Jـ*'Lՙu6Wle#oҐu *=G^Z\jZrUUw察ɂvǑ;R~1vtQ29*U"|"G+:#@ZT#9ے#gqY8hYbH2YdYo4'`M%mCmaޯT/aS}Z u|k^z+ QT֨-zWVVQ2hbz10Q9CZMNat&λboI@z>hؒfyijk0`7{߼Q|CkIǐ Fǰ7!["1d+u ߡԚxU_.Nϸ>] KnK^1 dlɧƇ4I'%}eLxaV(1a]'<׏wˊF"~ۭ|T(?cybs<:g5ZzwJxwpF9IDuV--!Fq]Ev wjs2%Xr蔐uP(![)!Y)=RBo^7FؽLers eՕբ_7">Z_>!RW#ɿ7d6],O6ϨqVѝ ؏(?*y^jy[Y)_EҼRA.CAiq(VnֵQݠ4Z p8lC+Vۡ[Wulk[ኬFz?pe8g`\}0슭gWlW#\CN9tWd<bs_!G\O:a+͸99 ". \`Z3vW<b;"s= \jsWlFD0ՁÕ5=< Vo=wDs_5ؾ WOּMӬD޹6rci~YMbf1]I7*V[hH;=tq-*Y;VQE9xJ_i\R)k !B`l*\[/{\Y.P+N` ˾Mz9Cb :{ +>^U0޸`&ض73֛S5yZW!zJ͋i怖YMVy^MsEJW4BQ`_S˪;%O7T'R*]eDxk`C0X+di׉Fڀqv-7Un KE`#L2f/bDiU6{/&  NR"WDXrEe@RLq+,pIEV(\]\iF'$W]2rE"WHx Qvvru9re<&le2r%"Z <ʕ TJvV, \-D?Ä(#[4rh2\Ng.dbWDDr,Wߌ\=PJ}&=3؏RE&W\,WV=0\0dpIE@rEf@bW-OǺ"\)S+U.v"J\]\)|sN\;4إʹ3>Xk xE|>077z]+/k XiH&\S1Vs)L2—h |m *W| }=\!."ZQ*JS%$W 4Iƺ"Zb+rJT.%BD˕`.Y'H+tA?J*ʕcO:VI\.$h5]Ir(WV1R+(yDB\nx*rBG?)c; =i)P27IF%rJs؋rȕسŐ)kُlչxuL~"DYz\q-ٷ~ ^ً} );TeY\'s@K;E9=J;Gk(m.~JFݴ0!m~CT-p- ۙ>CRڿMY^,tS/AhE(p5n{}ŵJ/r^1(k {2[U:WkfkYWP5VNAҖ0?e@Hãi<ŏocl2n3؝xz7lܟCgfwph V=|[LG#g5j2Ƈܪ ~w)/Vngq:Tw>{~eFW] ?>n_GNnc{~Rv?LLdi fC=O&d}lq y](0S.,F3Y0C3 0?7wd>͗7ܨU%[_^,a[.Y.ydr3|Z X:L qڈmHʁ IXסc%jPJVq.wCSni=CsW쯒}o7^0ofc4uPfEI;ww!ϓCi|JƎnRޢ,Wo0JkE(}[ &B֚5kXq>+ ^iGZsu,>}: @@ͦ;aZbZ5UF+veðCׁ1۔eT3Q2U`ޣPXT4L>a--w!%pY֕/K kg.Nz4)iޝgOzK2UA9`hԺRp]-PԾr6,-mv*v{'[~t\Ow`oyv3A76rs_-z:L~k4 oۨ}zwaDUB(tx 6jY8Y2ƘwC@?qb{9vc 3lVe9a`CE lSs5>hۖC[ك_Z=Jfd'\e GAgvJd-$1kOAVByqVE*-:?} hRl$; MG]:K1-.znx%^8[<#A_9HOEM%u6{} )t1_g6Eigo=2ܴ-,n't{乭InA̺w>յ!E[̰]qɭy>k6UC쉡nOe9]&ȃf!,0ejg*/g s'a>==[LS_)cDj dukua p`Ū5H\ĝ9w{H܁,Ǐq4LYTZx+crA24NYkߔe8j$N(e5H[0䶬odm.D@mtŬ sG I$n>/Q9o߾O=w@q侺W_-lf=f|UzcǟҲeF4^5Qc*XW*`Kueəo4gJ4Lr(PW#-q[2𲱕M!TNT 1Z+{y}+ۯȞ #AW P@G.^^*ۏ|K/KFe8gA,]cՙ>M0YՄs(;t8^h_Y'侄;8tYIe2`di]m2)UuJxS;VÑƣm4"{q+}0D2FmV_eL.bg-yB^κ}K6pZpd&@+NZdArem@BĂ245xN[jr@S(M5Rו0:cl*0V@wX1rpnG biFm`@{`p}c۠~Ӱ\.&Yu}?o l*9TA- R kVlZIL0n)s?K-rk?дOGjE˚6ZK[[/Mtn4Q9QI-Ȫd2#GHLUx׬FN5b- UCs/6nTəjvLJ9V\KWɛU&B& _ͷW;C /[ hf4\i51 VqvjCD5fyln-Ƌl7&qkLCD;xdV|RCwQwє-MwbK.i8#No`TYP֚Vwٺ~Y^rq&0G iFon"1|lyΑx2r} ~4^1ke3)}]ӈ]i6 -O%{=z~3bؽ8Uk]k,CfYŹ2R&>?Nfm%yL򠨍&K?{5l8^^ON?~O˿ޟ?rN?Oi=E%v Z}5 <-VHEk 2n?_~>0^\7laي qe%jc@Ϗ#Ԛ[O-zԎ\ 3xw^/U!N|.k=/mպ9ϱ Ivm$W('9yq10%f zoP} $R=mv6KJϿpX+g m3] 8 K߅, %quD2FTڨ9F{-4wbSLgӯdߝ5Jg ^{T9y,lU}˄8vn,jW )hYl+Yw $8rf^(u<{Cqdw0z4wA)!X@Ȳ)8Jv}@̨LNA/%6Vx D ֚pGi3܀< Y)ʹo<>%;B{GI B8#F5sf!c֩`8f]$ =O;~l BrXU.X|R r.G 2'!F尝^W='UM1֍CKM `EhQV юE0x*!ZXdl={Y{Y|rmA[^Π2@E|M; ƾD\_KB]ADTZoDoQybrNv]tj,lZеlKQ%gb}H?;&N-s,Zs$ 7:gǕt=2pC"0vzkmq)DFv%&Pc9T,H<咽BE x6bр10u>Rxp,up:ZjHFSvSXR$*E]4 H6-NI#WX:^X!,Kv~Mr*#VZ Lx9($ƨ/CMsԼU_iZҐ ޴%槓٢+akayPgs:Ghywsa1S Q!7L"Ao$1Xi}]cN/YJ>vQ)x}P~U>k:~ZUͺwXU+^Q=~ht_{1ek1z 11۽#K'`!C̖ wXkW4eǴh{LMK5. Λ@6k;U>`ƠAEolT`@ѰLՠ|ol1ct9쭣zeGN5 d~'%]rh`s": \9TfWԚ 2C>$dɥQȁ{\dyB]jsY%]x3'Μ>vsz3TC`;a @7qx~Ҿr9*?4*@LF7dڡJ =9R'PNs-LJ[Nt2RiAB{^D $3D<,J= IC747 =MdxnzַJP[/lG1_b*ϥ1'_FJ]F6 cyTNFIfz` ]O>_M +xeϠ>f!Kgxr3c `p7f#LHPu+-yr@C)u _Z>FC؎{7,O./'Q;G_n[o}-WgMOޢuI&֗NZD9 gmJpIJQ !X@m / CO190FsMR^)\*y )8[kUYl"`V)hS +^@^jE)rP<~j<"f z ]wNFW˩˜|<\enhKl0EwSDW{5c=/Ma 1ͿqŃZ;OIkL2: ɣ @-@E2O'H$Gb=~ Ï,~\.6pq~ FdO&M6׳8ÓPfXv2ì|򅟔fkLiϙH[V;mxJLp"S;jw8˺Hl(|@mE.O,fm4m#<5OoULL^=*1JIs%P1p9Zn EɴwWA߈TAˍ0Xf 6 V?ޛYl.xFB!tuW`W4`|ٱ'_~E.ݭ_ILO;n#7c3iL?ϹCB{2qb̯tIZ\GсTŕ1hentx9J 0 v+{uIa8an`t>NxW# pd7o((\hm&lx2ul_sf;gx7fyee rH"jY kI ådzKKqB;ƾ8Je߀v*Airɂ2Fc tL\`rQ{WE+%ëƁ%7Sj l늸YRc(ʏ6Y\+b3qj"meRn13\0ZǸƁiWZAk>IL\Q8DKb^ X2c]iI}7W7?y,$dq s^Ȕbd&Hcf_KEKuD]yuBuf~# 9ܫ޵s k5>9rSC3L Sήؗot0䢍\nvwf믦CJMi#&"WN{me!``3hCarZ&7g3b`NT`= ]*؄$~2pUȤjkj<ajg ] _.mqYb] v~5|_Єhh8_|C4'˔b&h!% lKd{U lWYc(^@$Mil4 B+<)YʛSfѺmLjƆ7L̮hjz#}OvI6lNL&#J"yAsB$&!+(@2䊬(%]E#pp$EE2 >F5'E1vE#VC5}{]a\RyK>ѹm4:f΋0 BY7;Θl㤕>K*dV&KJPVUUeX5^5^s%DJz^/蒶.J pޕ$ٿBm]vdF%i`iN4jJ5C($je,^fFx{>D)22Y "IuF^|^>:6C3(쑂"%m|t5paoF&@7f?(6px{rŚM;67⧳Cԕƙhb"Jx3m5}{|ؠWO3> @I"3l`Xr*cRe&W#\%axouϑ. ."HTɻ*{ԫz+`Z{GČzu7zU^WJvxVEĢo9^LSYIy=ej=?)Z糋;`~k͊_vh%=/ۡ[EL2ľTw,lL8~=_Qϙ 0dߒ /+-$u[^-D{KW+d0gvO~~vm2; ,jdn_o0'-7S_[4)L{ +dHo^O'/nu'Ӌ\ѿzͅܜT۔n;ZQ;$5Iu!H@9)Ϲ~6q͛lsC,K}̦:VZ6NQTJS+T8J2gÓRK6$LM.ɫ f!~?jk 1隸K6# =1L a3NwQopGG_ܢ,JxrRyGsz)?%+sMoosv(WYO-ˇWR=QD)BDPJp{o2sRB^y/YefИL%"3 ;%69d=s۫baYȦszsUk{]dy|6z#=Z!Q cn܎rԚzO6Y 27_h M1.6J.=d*yk"P GMDfrnNXI(yfmu˞axF(Aj>>g"S=GLBJu üK(|f5ϤB\E I-0ET&efW >gP}Ho*_vBy⨣yŽ݈>޶It ̮%Xpю$lΠyq"d~m,hG7Z4[v*hGGdm9$Yfdu1Xheb45{G]a6i=V[ʵ$ݺOp│p1'CΖhED>G"D :,QGy;[N: V/y Qwf9q~ Y'7Wy1Wݯ/ 955J$嚷~DGLČ& * EpMP5EZ*N8f:~gA]byypw vu;'T+EʌtЋ] F5pY8 *ZA7ҕT9ʟpi6yz5hqL<88ϝĿ=>Wy]G[5fe~WWBQEj v%%{Gf\}>BW'y~y4skN]bK?pAqD yzDDSw$2[A1lC.%S ,BH3h*;I^mNKlAZ=$ǀpQEUZ(8[=DEYx̱_k`v[M3` r04,4]QiiZT$R.`0ZmhQZC#] ])ZJ9]ZC -3FJ;*`JC+FF;]1JGuutejF̀* [UE*J;AҕR!ѕ8 ^&1.]DZ֬7ZYL˹LOf׼|<9=-f&eܞ|;OSnڛA~-e;WN5]9Ci{6|=flyw|mbz# ޏ'gUmW6')Z%e"8Qtrm qBX|h :(A! $H.ɕ6d# 2>$cBY/L}ˢ̩"6htAk U Z($5KBbuZ(^ЧR)hnVx%\yX  u`QmHg~>__6xN裐Q!D- lY \oc%1Zy^ i|d%ыģ 5\q"x "ZX+)pzm NEo=8u羖==H:Y*i,8D-!!֗"l:eRdX$:? -׸1Fx+"!I(3$Z-/hg{d &VڧN҅;מ "9:[r4)stIAێ& ﲢ1.K̔}%8 0Fk-S˚u k;[qX%^%O!'*mUSJe wy@ pLM}:CXvVUqՔG+:%q32f0 AP ;2Y0mCUlFL*ωy = _ ,/OyZJ\%[ f,]rG_֎qSu0B1csXaG^G"hIAdKD 9~N̮ݬ4 3q+(\|8&X( ꛐI:Ci*[PN+$_PcYTB2V^X {zeY r3Jր1ucP VCH(PED&TD{#]ozg-JC(ʨ[sV ,=\i74a?%ӌ/%RC%:&|B(67L:A@GJ Ltf+%RI9n#XT5oFY#<;"JPAH_($[m (ĩ(Hv\F*^eFTcܣ.RVK>#iBիf !1&伖H lJ1&dY\D d`niXEjE},BZI །rIڛŸKQEYo"棊1͋BIJBm*d]9{Xˎښ]8%!Qv`c{2]` mӇNc[oT^MJ]t4CҕjJUe ,äc*Jr z_fT(3ngל HDNW2^i C.nk؀M;Jꐎ:F& h%A J"-{Tjj5t뭊UFd5fIX@iT_}ŷiz &ƐZJmm˛zvn ndeqڵ7`t ivù*d}U ԭnJ''f= ]Z`wwk'QEŪcԶaZSQt5(U_$056v('g 3ܼ62#I*pPaSDlsE6G=Tʍڪ+-hGݔr%5 2TA2 5X@J'2(;HqXoܬGŰ%6vł*'+ IISL1^MA53" +U@ &eCE5FHmr3u#;l 9yT,+~x$mDVc-S{N낝٬E5jփ*Mg(jtϤyfj)څ'е3qTkBT>@MtVkAW YcV9ڢb7>jNàe+J ̀|'zʤ辐fhJ7a#ȍ5>dN֞ +JOQ!,J S\Fnj5q3XoyFmԔW  "6b9\Y 5I<,* i3uc'L\- 5uhD;S`]pB;6WGwQ P3l{J~AtK}@b> }@b> }@b> }@b> }@b> }@btnJ`d|@WP~Ac> }@b> }@b> }@b> }@b> }@b> P}@89`{~2> K7}@nV}@b> }@b> }@b> }@b> }@b> }@b[=pehM{ll|@qKP<Bzb> }@b> }@b> }@b> }@b> }@b> }^>]+)>gtԔF=oG믎ӿ VS2uw5F?v#ۖvkz\nhn(/Rٶtߥҙ]`&CW7zn(e:@QWE ]Zm5LWHWJ굧\" @]H9AY߾ir~y%U5?{]GϏ.^Y/sbRk$т>G/Av2[96W1_/EZXw?ٗyŗ޹ST&.ب9_Jr5ysҋvNpĪ0Y?ռS&omc`ӿ]|gӶa/O8!7ìiVL㼳? o_^>4XYiٺjSa2,<}ڂz~݉Z ܭQ{[un/g'V?/׺v?Ahg%Of58 z`WD~mvvXluQ.fsoy 6l1d"<3,=MnB(m6K؍΀>m}ZOOiYZos1:4 ~ÝLҋ iem[G|wc}9@o?.g1X媽#ߣfI~>}1yxh|-5ΝCam5+(zxFF Za9~%v׬ͷAK)`*Çʃa Zhzaܫ^У|YP|-%F LV@/?V0 yz{ \Fص#+7ݾOjLj#>F`mwbw0 =38NtrFКw1J'dcp+ Lc*BWVGPZtutN?]9Dj+'shNWrT+ЕWޙ)+@EWWMbPztutE:rBtE?EWNwEh LWHWQ;pr2tEp ]ZvK/BM+TJ#ܫV;է+elPjBtE ] BW@Z'zBmS̯ݓ8{ }yIˢ |]_a߾;'B{y7CVYK~Vymx2hmq]At٥[|J5o-Qq_Ig3鑄JUu:HmnBГWMHhwK(=C"NP+TFtZZ+*]N^L-Փ+BH(C+GM0!"^MnS+rOJtutd]y-[k&ӻ"^;]3 G9}T3h޷ euutWaJwXk ]Gy( >[.<{л LW;}duHtʵg]J2]wƸ [a'CWW5aP:tut N"^y}w\n |gw_[\Y؇~ӗ:zl#cVlS2[RTR)aA*cޮoh^=7#}~4G,hW._̾L/{z Fs7 W_t3{O߿9?НUoc F`=Eߥk'K_,"}轐eoCoxP!B77iz׈k= &ˏB|un"3`&SOKP5A}F \`5Bo\kYOvPl;Ըj%]Sa᪝Jݳ@= bH P& Zw\JJـ3>#\`M6B* n+TԀsĕq aCʗhWPC4Uf$ 㲵*dX g'4cH)|O0yaFh&joA2x$٘(׊\LaP+)*L34<'\`J\*w\ʾ7Up%Tf+Ws5"RWg+%L3EqǺB'N`Uq%*''9lY6\pjiq*pu2R0 ]\WWg++9` ʵLcSko\}đE/."R޺j%Xܺj'x0N8юv*ep%ZJ znSW X P. Jw\Jm\!0>hi1E.&@8'ʁ OzCw*N2N (%{p+Jl?kY.C\ h*lIQ6"fD I8F)zpwT|+ X+ Q5`cp1wLJ5`1-4f+,YAZ+Tk{oU CZ= qd+5$\\ վΗ `puR2`|PԹ jw\J3XW+m%9yhɃ E.W :G\# T;]JKsĕ5d2iF(f㿋jYq*JYXW[ 6'?\sb\Rk_:!\S{vnl+9EO=8ט W(\pjMlWg+&UhGE|I:2I+}ӪPc{E5 GF&!Dfc\ wU*3ghrbe{OUP2*ơS-mOo+`yeiu.cU26` 1- 6F̄W(W\pj;@%'vJW ذ|qQN2|P=Pec_׀&Zr@b QƺBpu23 "\2frw\J9]#,TO+3~%ܮprO}|F;FW(>*w\#^]=uJĸj'<۩'ZԴgR-p\=)#p_#VWB Zw\Jշ^Wqك­ +L.BWo z\m.#wT9լaI.:gi,i4xR-XUw~͝FK(9xAUUYFPP߸֟e|q;7^#5x>Ā}w7c|͋P@g}JVt fɏ+(˝IP. v|UEi m.|搠M/CT xGo]oU7s~r7ChY-|B|~3i{r0|"oIk)/{EPnѰ}F7󊀛0c~b6K_vHSg\<"*v;v J,5j:^L./ T"_ph2i!y WNZfP@)( asz|sqw잾+O2c\V\18?+Ǥ09xTJp\\0UPfs%gx&q FnO,#K~UղE+:NY겞W`TWȶ&aWos?|G:7>Yͮ7n}KgΦ~PB}\mZZ~^gSkq>͆U9Mo^px[W|YF^ypCd2$¶.X.QFcD:jPZ$4ZtNzF ("D!eȃ"D0 $P/pRbLr$,P-MҤb,`!kDS0P- BB'rшsj֜Ih);OR8+.Й0OK>\=Tg9Ǔ~X\oH3ojS'JiHlJixP΃>٨Ds-!! 9 ( etz?,V8 z|mB>( x S$uD%ႀ(j%cR0@Nȃ 37t'&Gq"+"d! P aA)'*j'IBNr:ݵ>5zݓF_5ևŁ6{хSZE]ip#";86M'"|H}8F,$*ĨXX.x))DRZ(!̪ha5Ovx{W+iu=-y_oSꅳ >*iyD+U*f)ѿҋʉX-GL }\yX]UyY8]喤$n~ϼ \PM%PE/Eqog#IXNʠuv2,bX?GoP(w Y7_}5a\.? x ”/o% l1m }N62-fZ=}?dR.eQVRudEF~}U*GpLq4>^9u_ᅤbDL5͛O Lojoc*l}Qe7V$>LZ);C)YRQ ´ň6zW eK~\=o=hO6ѽX!pcMy|GtQy}w72O>E1mey/EW7zǸ-gʕWX,,M.?ѫEz, +:n/isf>]/ ȋ)za gcxDfڸyDyn+r5`({s K O8]=6w=]wO$&E˺J3wȃ$Zםn'W{|0_͛O}Ӱh9 P(=/*GXbAl(2ZJT3E9/``]:>/'$&|}~B-u9`Û=Բ eݣ껚5k|;(Egq_2uWA :$*Ӂy0S)% Lнwgu79rv%]/a(ZcZ^Ӭ. nj˱mwp,~~> 0c̣E7JqrZ}Lj0t݃\=|=]]eib*$&7@wHj^мܽYKI ge35]3d$vhW(év]Znl 9r:_^cb*F'!ʹx9s4haͺ`ϴ TRAX2cGOFh_ {3-&y4lL?gj~DVamZ!R9'DSZÎ1ا}nS*}¼ެ&*̩H2ʐX`O)ur(\KמByvpu鵶f7%XUףL9wF6ݗ8s>>nƮrb %.cFL$R"Fomjٿ,/s0n3n?h*3Έ:->`zyWON>« %%&.CtyYQm(~pkEM؏f=]`O֬Rܱm>\g: 1.^V*4d<p:%hTQ) ݞ6.b,'@]zMRz]()BR)\Ꚍy*\YpAT}Zn?lyW<n6[|MWNl% tQ 4(ǃ$Njgzؕ_ h;ܷ{s"( v47 =ŞE5Fv 0,MskX K*iW rȊ,Cv]5[wLxβ*F"^V(Z)X]߾ Շ%Ck Cċ.bUa%p%`Iq9ԯRb<؍?ՈFLF4S㊗jѐ((>x}Z7qWEGI щoYj@3LY-Ci=vvvֈٮ?gyY|Ռbݸd/I/ߩgҋ^%-AS,Z$;bR!h_F B&iҋϡEv}~Tؖ |௸"b! nJ~dmQ6v~M㓹`\.:; Ԑ2bB)y:樃g>;Pus'05NKTcl$,[Ƴ%&zE9>ƾIm#dͭ /hz”at: l {DQ/h=cg+v~û6lQ7F_5^>YT1^}2P#BOMiXԖ(@Umf <1! 6g*L6z3@}rkxx?e۞l%zOmp]Br򮺶`LlՄT)rX[S22nm Te@U|߭*%/btNsvC !'iS:f@h}7rvHz@p@;Ũb_NP!wzR)Î|h.n o\b°xfiG3OԌDlƌb4DкƟ*]&"FfL9F$$w0$ة.ZjεZ%E89XR@W0^윯y~抇6NA^wcCģ,!a<`E傦CARx]JVQ F&'k0vPwqȾѹv7Zts,%*dN-#h')䅒M|mh|rWG?3+ʊØ޵WWS,*@mRd%->UyY06=i2ARF]mVrJJ@g f?^9;?PH {@)~r r=W<-~je}:7d뼏58NБI+j|4r{-s>ڤhT; |8 gsِG?>^}+wzgr}XCDɗ| r`Q|Mf,ٰɦBl[_ WDZl#rgTxպ >\a)k|;m}Dg?}G*1]Zj&9dV&s *d uы+._$1Ʊ!+-O2Y{(^P|;c^(!9ݿYoWGtşi x !&n=k^u+@/rXդkLEo1* \R#0-bȬQd]lT\F@q@l|ڄL1JIQVqZRe@SvΠ~|b'@) HAYGGHU J;g]ZA4sUP)ZɂR!JZ@P|+ٍ /eCYe'_|ޢZ8X'!.UPٓW1,.y_qqCC=s/ָ"<>G|Ry vgJGlbja C6T[D}pИQqs׋fNnG+d?emȫR!S:EaqЉ&qhCH×2Ne W1=Cb"r0g?Kڀyȃ1X)UkXz8q{US2PH=։o\_!¢x!6lV:cnFzv?D!`&_lX|v~) \5#` cg_[b} zQ3L\QZ N[\8YQ1Pܖ'k5\|<V$Fγ3X+gYp@{^8K <&ig[{=yW=I~"G ՗8N1ˎs/+;|#KnhVY;߄["$mަk)R%6S~Wj6Ճ■>x9'_6wV>WOou{ޅE!w_rⅡ{X<k\j ' nF4Lj-cKBpw0S"m?M {KusK Ͼ^}bV4~XCh,C(k=EH܁L47D$N$L`x@Hs2Zo635k 44xt3dK!8 ř|u\LRL֙P[gXΰH9׍ yh˫_j~vr]^|\m7 >B\_&>^BKR9! BaԔV!xFo*)O{nk)p: |AT3\:Fc| EjQdt#g?~9n{T`>jk)fb,_ȳ&5 чWӨY_KF ިQԨ7l,eaAEnfH CV9>sbp1nfGth/Z~nfʕ袯!oUjS zQR< jQ>hOa[:8i]6LQ-<$ Ba,hلh 9"i uJ&"5(OA/诸?SaM{P((t˓奯 oHxp!kyl>aѦ֓u=:h-,.A59ԚЛщ\zY1ɹ+N;35 3zM/<~.6Cn|Wg]cZN).Cl-:ѦYW}2Wmu|0ӛ@V| V, TL lbAܛ6CW!9_DH<AE9/BSDžu5:j|֮*j4GgkF\ɄA(.s:++KJCcE~hګd3T`IN7 L&mJWk';Y yxARUa֢:y񋣥 |A+_7_wyf2͛o:v|=ggiN[ HG}hc(AbH˧_̯E4,Ri,JqFV)dk8akr<;w~g >n B|~@b"dQ@ueq|XRȱ4yQtfo7SuYgVWɋkAIJuNo'}W2MqNVWϖ5آoo|W`]au~,ք˱92;:><0ԣ?RZ%׷tol檱|E[珏~6͝CU>E'7dug=^ް4~{Vgmi+zFeGKGzoR~e۳*r.ӓg0, 3XM-5fWfpvy.oh/k/C>, cq`qAqXэnh,b$_t.*X --RjUޕ\1RI#4@=z-cOpA.bKn]Mܱ]$I%6\{Tll2ۂ?ɹuK˙ Sڒ${:٩ĝlb;yT:y]YONڎb\n /j#%\*VdH uRnK<g )QRgId%] Jd&&5n"OV- RYyr#9P./e)c8vlk4cY]W'һD` e_jHSW8:,ir-U!_la4xۇdRvgwCޢbC.p$*L(兣Ҭ ZM½DI85k[UbUEN},J(a7h>g7X,b֝ETU_m7쭖c!rr-ۡ5vZ7:o kItͿkz͠vu^=V9>z?n=KC1^Ǫ?3168?naOпf=S?c}*F=q#5UծTqeKN^ڻzq*{ٿĎ{v\dEƪ43S6V 6&Ėɮ1z&9=~(ucH?){1Aa8b0>q~rMl]~7#19AJCIgNa5a}*d+eAܳG𺟇XbQZB]5Xe[: lfGX sSn4M4Bcr'aZeJD$7w^G9U,KKps[je&T f2:\R vKA!/KAy# dÝ]uOW@i ҕOPr<@QWS֙]RԞ ]6l2\Kn*tѲu(mgCWS)vOWwͶ]ZP^X;AWttuߩ' 'CW@k (ݞ ]i OI]0?p ]u>:]u{uD*4зeŴQIǿR2Iuy3 2ccU4n|hy<ay|k;f<%/|z>Zw:W%ZX:/gg?ON*E~Vy4+kLx6ؿ_-唆ب2oy":4GI6LI~I~gAg$7`IK !5?RWsd]}(Xݴ̕\V`^\I[$ƭk,z5ϕfŗ3wCX}gSxf<1O߿-'$;`1=ZH֘]۽~:R'DW씚 ]uz2tڝ/Q}:/d2t3p(+oj pd誣uv骣'HWyQ4!'wAUG޻tfOWO+)]u~: z*tv^]uWҺtt*x&:`?uɨYiu({W]-^)o_] Vl+DKd;ЕSO>&DWVOj ]uv:tJ[kͧz}?;>Y0 bsl^yMM=W2L9Im,6PY!NuU4ˢOp˩ϔ:l;?y7aˏ V0x7 ;Uս34%"#Lx2tEl=]= z33oV:IvXifٿ`es"g$7 g?|?(olA=Q)ڠnY?%]J_~xUt VM,&z_q1N䓷xQANpv~| [Ι[pѳ%E9|nT\ܷj/^ ٚg/0ċyZ|~_5>u+,lrDT*@#(?>0{ ˆU:gCy&=&}0?DGr񷞵})ij:Wd u&j랧Pfi1JĔhD.vFJ86'\hoW4c+{i}=?kvZ˛xށ/[3a~Tp9:"e5WWi7{BF/qrRT Z1:}eվN11I5 CH1bQ-6|L\nUq1߷;MMVtҒ~gKa_*00:P(XRˆ.S0hVR!;ІmgE kk֡tԔRTaVm|DRh*kbiM|ˡޡfJj!(` b7ؠߓO nxt=mّY+O`ɐu^4>Vys#*W55vkxNU`U3ATs$enu})Ș|نэ@ۉ(5%bík8+DLH*nBYZr)R<@*=%BBՕm]*"Ux#X'rdFlUQ(c Y\U>mQ2!– 3HX6U.QG5 =V8I_*KI2i=e»P4>CZ Xv?5&X EA^ʶB m>)h-S0mLݬ8_UƏj%ʑS-*j]reA_ <ʍ!Kicm5eiX,ąd 2VCl(4.u+TRM Ap lSfa)-%Տd*P:d4xkkAL-716x XfQ\5,%Vu0 e ldW21G"S(R3 JS㷜2b $_X:™lKF ŌM J+KX;.QafH57]Rs*Aa ²Z"01v@ m7lVHn:!nMI{$FX70a( }̕}=!P:,\ f_ ` ɄFB8AP{s0ÙRPpq ,s8e`ߢudYyw= dop_V* v .+W1j{TQA} fRToʐ_N7@yD0PF 6Bl A$ ٻ6,W}3-K@d!`gՎֲ葨$"}%h}(y֩s ӰI&0C,%D;M[ ό>;a/xr]WWJ,4g-(*S*AevNZ#'$ZCxgWrf~8M.x Q: !jh!&F:7ŃKy@r *}ndϫmI[ sV 30)Od9!/+X{BzP`2Li=a5 C;C `xKR@x y- mL ukmo<& ;`f6े" {USQy$*ƝSL l >gB^V2!HbGkn~^޸'sQ :m֙Xk[Ȉh0RKnw {s 7DBNyr WH]֢Cg7 DHʠv0FupmNmvgskay "!zbwP^5k18X7AБC΃:*L"(w!x݆`1^IboIx9 }A!zqD##"LF?{a1vzPB!tшԠ0 [9Z g/h<V ^,ngaMӪѠdfeTJ}OvW%d;BZ&t@& W,/Ƿh5C/ ]`z[j )۳fty^n9sf:[0 `3 *K ׳Z nUY:Z nk ?kmcf-Vcݷ ƴw~7\}ʌ4&:r7 J$𐗘K$u}9 ]"HŹC^%fn3f*!CBJ@29uHπOt46 b^1D6~Є<Hw&r;[;L{]!#{tF. 1e -u(NFR!x { \O` \SclFu&{*mW ڌ)ڭ CY QLJ-ɔ 2\Py tm6w[~kPq%Ԗ9 }C  VAPʛtLGjڵf|1HBjqc:zg,^sOf{6CXrJ=Itk 77"@.2 v .Fmvق˥11D 0r]ةPsƛt.Xvߓn\[@, +]Ot"BQ}#tG}`AAp]eߟ˻Rn\mJ׵ae* (|hAMõ ƶtnHN߿~,i4OCvtDI7tM௿9u ,pe>.|#)6veW 'ܞq=ϥq|4ff9t!|2o5N]^pW'^i~Ih~W.+\^꒯:ܾOtKvDzZԞ붞Vɟiܡ*_u81 ^&H|@$> H|@$> H|@$> H|@$> H|@$> H|@BjU(Ej[Z(>0@H|@$> H|@$> H|@$> H|@$> H|@$> H|@4SV1y/ Wbb4yV|@s$> H|@$> H|@$> H|@$> H|@$> H|@$> h> 9bj77j|@@}@&4G)_|@$> H|@$> H|@$> H|@$> H|@$> H|@$>/t>=^s.5Vviu9?ݹrSu˫3, j-a _mLЦ4yYHlKs- yU5t4ր2nSѕsۮZq[e{w_1aɊ_cܨ"^Xby@[9Kn{ھ=>͗n^?Drip 5gE.Z[om3:珋w>kq{s˯u|Qn+!Q(u`i&GM M\)[;*xqe')9m㜐XkCVv! OSgV ~~Zt-A,Z`X/.l&z3)ם8{ -^׷\\,W3g\1V'Nuv^Y>v;lM[Ha%ck@\`HJar68[]p,42\k+F'OW2.G5-42T]ը+F;}u(!]aG++n\+%gNW@!]ocxF[ubaU'2I28GP&bb+NW8U鳯V앮8bоgjt(}1t^ca`Rt5>lMHtޑ(ڵz<]m+tׁ& n=thIM BW3+[YF/g*~RgMtѐɥmοkWF*Q~vC!)JM 7 ӅIgWg`Zeuq/{\?yNz q8}L:Mq4H7mJSBT _c!V#nF~*4J &nަ玊XM:NNٔrGiw*hT M3hji5*LB3i6T]j")j+Fkit(!]N =CW Z hb$I'+R5tV:BW&u5G *$RG #F *&Ӭ"b]YZ&uЕ~dcQ;8haҪixBWOzH)$cT ]1z4u[j>teTpFWDW 8ԣnZ hv>vBWs+TS>b\cV8fjwev3ʹ׳dl}%K5-Y>U# 78]$dO]2ʉm(e$6F=IԺUXp҉K(?؝Ƿ*VD ճpC5 @䟋0ʩ[_'Aԡ 'AD (BW3+o,隞p8x8.6Hnʩ ]]IzS$N9U4Pؚz֮.UP9U6Zp7գR;tBW_]Gv9V6Xwdp`qp|ZH{?]t5eXMAWF]I ԛl7'6PZfHWLp5GT ]1]{|ʤIjt޲lL .T ]1Z3LFiHW1D5=MtdP ]1Z7yu(Cbt̃)~}h>.2;Гq(Ğ "vV.ZW]1ડ+C-tIaBW/BWFj+}5tp#BW\'OW*9ҕsak`2GjA|1O7Gea*)3;yzz<Yt*-F֪5kY\c45F%[&wc58xZ7K$4bw4M4uJ=ݠ_IcV.Hy A/%_Dw˳+k(`iқ*Feɸc)5C7|_3/6|w'w܎uq]^Z-- NK^-5?`x{z (ӛdrwV$wNk6u&ѲP]_1]ݿ]qi=S⎗wKwPpt,Rr!W5?q .>^|!vzfph`kG/{lFV=^dN֧ N驫5F`ej(TDW fQp+FIAjtWd*CW gwCK+FD]͑)֔ F+ ?{WFO3w6dŀ?%dg\F՝b[I${~U-XӴ, 0j=*>|RE/3NLWCW^I8oJWa: A*tEh=DOW2TLWt= >74W8*-UJ! +tX׃hlBtEw_1 >v"s9ЕX]`M2tpNj QtF/r }#<#m1ŋ;we'G~c.RݛɢJ;VСD0fmVݚ솄0ƝɆHaBk_hdd)RX]?]1`u BZ=]Jj  v'CW WTZ;]1J>ҕAIMG]1\ ]Z/et(tte{_+ "m\++F^P|d0:3 J|?]1\/S+Bk9]1.Pt,tr>Au:A!9F!9jP!lCW͡0 v~)ApqTZ'vFL]LWu=jS+lLTZ;]J'T=+>l2;-Lzk,{3 ge#szIܗЎE6!ǀOF1]B+%.rYSEQRKp2Bk?>˳/Kz`3w4ܹDC%q:;8U}mKӂAJ<+`3|Yt~X&se]XG^tM\ ki6:ĕ:z}1BMK  ,Z&FkRэVu#>=ԍI8mg6r׹B/kUD c갤Hp:4}D `AHB,lW9Ŕurߴ&~8ϼPRVJgYWo,xM<<-yO0IA:6&'u &M WTd5"v(mmK[]w"Flިh#ͺEfg7fPR[+oMKNTZxx-דb{ FKm( ֯5UA+iz!u@ݗ1rUA0tp8-[WTl.>+,&rg^#i Pݶ;Fo% b1fr9sKe&h->c.\Ǚ*'7pU2U*t2ҕ{ntŀJOfy-0zu(9+@HO'$SĄƟ(u{9te{R~kWB:y* oc C`J]z)NITJr΂D}{rƬ]oV*K '#TnOKJDBl\\ V +W$d*`Q*0NXA@c5r^q |R3wM24ԩ4E4DiziA;Ҿ~~}*bZBWvWԇ2ҕ"G5 W§BWVE/(uG\\'DW غdz ]Z >vbyn/I#Ș Ogj5&EjKe;9nelDWVEeDWݡP vvwah͎pmЕtX׃RVbBtŀK ]1]03]!]IE. }ޭj;]1e4P\IFdt|^/(ԗZuV8jT}a|lb5jˆ-Jh/gj<H` B)4tljrD E )t%|1ġ/rXi9GxO7q'8PkShemY"iGE%M7@cǂ׀GliVS`-L2B2kVa쪗QvyVz:!b^&CW ]1Z~MQkFth% j壌~0c:E"<:ՁȯlKl:7В1ZAB{$$$d:2Y>$T`]g") :jRK%[1`qLB4̀w^7 s4";iziZKH0qu2tpU2U%Dj9\-jƮ'ǫ^V?W?~=8OOY;Mwg__O)r=dѝ zO ᰩLq\ 𢓓x7iA9+8~?xUV)!>'%?jMp(`<Tޔ`1 CS Q}-2|VէzR5_fꏞVV3LoֆKUo 2 _\/ԿJ8|kJf&sO:~Az? -jDqU~=]q] {ܷ鬸*g|e /M7 [*@;_[Vz?(-f@}!:#v֝rL??~}^ֽr!i*Ws "&zχ+o#7zˬyYAw \O<<|bsQ^<%9ܽ;XfYރRm˃Y<ْ@ٚ:nnsY(,yw췫55W=ĸ) $,BSZmSd2RlOtROCN;PO" ;n>Xߊ*/y|Ϛ-qT-~_<^X&,Nyoǽbc̏_Ztx7i嬤(R Gt%1ʲQ_jFB=m@WJq(\aofdg>'sIP]N8Tٝ1G9p,ݱzU 9v§0cx\JN7_%!&:]擼s>Z|џxtvIo'OsAGF4xQ7b(Am̰VհrX9a\7ѕ5X_kQKSٴ-& ]oVJa%|1$X#J&w_ʞa*}ᓙcE|ٳɶA3Yr:caD*5̮N>}rFs13g}.df=HgE50Yf)IRyYfŞ;5YJDYgùr&Ed3&YMO.2)''"T0}HRyʼf=ηWU*5.'6=\k׹0x56e:}uҵO`2 Y筁4_H%1)WѬӊPmԋ6mGOM_7XK" 2)e=,Q k(Pz3~8YUzlK׹|=Ervy6H媜LُǶE[9x28\NG\y_~iN~$B,4(+[RHfhߴ-l87ob;KlnWlR zV̧p_翸9~ۣmdq`~vrzŨ5C>#yKGhZl8FZC_fV2lJ;pL׭wNjϺG7nۨ:%FVDe+Sj ES~(9ihqBtBqc?z|$Nn~ZS!ڸE*ӏYV6 QTĔvPo]|MtvoxnwO_o[u:LfPREceq+ NǶ|q>BOnuSMQRW)UMTTfA=+̍16k 'DlgFSΞET9uG.e1vvjFua]/سGZ_liU} ]I,Q@[sq96Iheս={jSvzQm+{O[* ܸGȰ7ٗ sdAU2PMqgڿrs$sü IJ-M f EN 1q  St4kgm*)jV3N3gV0SԬAk2QTZ=34 cDRt 7\ՅBijJ,Gh@V: HQAp *m%EFwku4ZzaXcg.qIB= Jk~ht;y,$JtG_JC%2E, L2~ojN9:oo6;|},;ߺdJX[>ǺZx=@fzjKkuބ 89i$ҋ8$XK^_w'&_)9iM/pBgY7ᵮ Ho0t2ձmA9/;ӣ6's꯸i^T_)`g:}|>ձfLb ,1cj\zKì8'M3}R87Y~KtӔꥱo#+,vj~f9-I~rCD$=UXTŸ泋$oEYJ)QTzK:H{[:VZTq6xcY1 6x䥏h.j;ǀA1%Z dk~绥=Xp5d3+ R__]-w;[kݴƁ$ K Is}e_4I $6h*泆84W2 ZK $ZW։FݿL&/^ᒆnM}S< bkBQBNE% ! GB ?AOGe˫!) )S)-taR Dă TPeF{P8֐|rЫ%ϭEoxd%ރpAx%Vїw[ [/u~ M ʀim`z ǎ-݉@7^|q=fW=aу5Nkhh@v_K" Fw+o"M fѱe"_ oċ^d2WS֝2Xu]xQ/D[41h^ΞzQ^-Ԕ?U/Z0D|&dp `%®t˧AEkCF<ʩ{\_howɬͲ5Y (B[^jB#?yFK/klhI@Y3&,M~V3W4WzY>,/[RQ0ܱ?0fᮈ*fZ:A)R%Dx|ǁ gAψhw6?@\)-3Bt%I_<.2J{6r$jTK$RBy0`Z<,o,:jV}|͹R)S@wQ4 ԏ|Fe$x-DI;||f }-O!<ZPjXXS'EՇ|1C/"M3S@^Z.|HTCaO1aGZ_g4 8€Lb0g0J{ Cﴌj儲*X!u"Fp"(?NpaagŁz T Z Ï\'Y=Hs$Q2Cb,F@p3cLo@y./,k@Dx%*¤[\7"JQy7Oo} 1!zg-q x4>$c7.pZ&ezH<BZ,èo+~Ie]>$-ёw{V?j5EZbؾ΢~"bzN(iU8!1=Mdz$[ 9\"1c:Pe{!Jjq!ff2V=q+yth4z|h}Xb4ЉTA金X"cPQiDT^iLi{yf6̦c[VWstRkzK՜1aN dYR!Xf)8_;o4$2_1ȈjVڥL&_(ѐ HA),ܥ"_I䌜I,rHDBRY."fe̥)40/42]Ϥ':X5 L,Z, s^Ϊ0Grs|BmSvQ'>Jur81XXS8Kկfva yL~@k>]xJ $[X KQ PlX#kwsQC.S:5`$]%bW\GjcDKuN\X5Xhkx9d}g!Y:r5D2n0~W0%#"{U,WB+eoHӄYoY31$fcM*L8̷ ije(1 ;6u\sK3z1LgB0Ns3#'ݵA/ѠSN;MI)v!ԧz;]vqZQɛw9:E:Y/Ў3m똗@*=4S0j~0<"O8I+72OօI3̗?\9I>69 OSأc󎀐Wm.f"+?WB>ʀaB]8Y7m?&qn~ż2pT*NSpFnқβZ-Bne[ E%: ]]nq$E i'Z@nO ߾I EVh2ӆ0]k}'FO;P YJ`Hkí~(0kť؊m߰l Λ1>^~P%:FnZ5.gwGdńn9(YP3s |[޼=<'2g\fc>+÷0Bttr*3q2u&DLop81eF%֧-]wt=_n 7v\Г<:a9]rkIƺ6O2-E>GR'^Ԥdwb)!*f$+η@]n=0)X1gXiয়ћ턼pƴб۷ N3;w5Z˳hkia!c8myEqƙjtNtB|̗.b98cO(x](hdb<;(J$WЋNaȮ @Rʏx 3N5s| Jc+?k/g^T2|}=.t#q\ɇ+?NfooCi_*kr/|/ da*LDKq~'L{Ny'_(.ɚ|7?,GaKodzWiӺ?[-XV -6ͦ͟ygUH6{gI@Rsk %==PkM;d<0dejc*/yʼn7F0hm@띾Oca{FILƳF K Z DS-n0^{L>yQ.xfюϭs!CvŇ$0j͟{hهC3/+Ú~`fR ߯^S߮HB򟵫Ez7:3wyik4GjK(vP/AP4|GPUa|+j4gybNtp D Lb4'E872F@[LUHxx;O]it!Gtzq1PMFCTpeNw鄆1CMPp;pe Ѯ굽,t(䐽Ñ&C15Z5ai8BXhTy3Ab&W~ @/-9Bplxpa@ӭaEƔ#vϴ!xQݕ=uEKisXԃQ%.P6{xhE.@Yzm%ř!B0%^kL1R_D|xЂĽhzBJ hz3{) P%ioЃA􈴤WjEMF+zǪgRa:{IQqƓ GjcP;EJˇ-h0jO; zҺa?`=Td^ ٻs$ȄZ>nC(Hۇ?H~\\{%i TW`ҸUET]_5~jK\< E ?ǃt#iM`]\+6ɘ"ʖWL2>T4vll#@TFMAd B_pfz(uZ8\^oW%@9`XxS[ {1,uW"^%ڼ*jDa$ew|oսGE&N}'CF20HcY8I oav5T5Y[xw7mRr)[ uKV= ]݇kuRu:@2U 4ZP%M❩7eK5?q<*+e~DO.Dܱd(#J <"!;fիRaDޕ47QV= Dx|x1BqkZdI R}pJ2LQp R:M/|}F+5L?>?v~f }J|E/oS^D9" j\F8'U9#;G) _@!GW\Q0T6齮:Y0ѐS6 1';F:LN<Ff$;s2ʧ|+-IF2#2>0 :luo35OXSkEF_i}R]JN~Rc FMilJe +%+, L("7jD ϭgHwbTVvŢw2 㷬Z{{7 [Mj;#YG19Ivv39j"LD g…FhZg1ވl2jGȬyư5sF]SԈ0Z1l%;t]@T#K8]g5IVEG%mzWG>A9KǰSXY^R,|!ݤ JdF*b- :ȝɣ?|:M%a߫A1B9[XĔiNR/R:&iVH*cc'_enއMzz!B>dрxfT]tj0켷CMSsY뛷xu{CL,kU\΃uE0Eb)*,󴐑hr^ !3Tti-zoap]I>R$ky9"{LoE$q/ 5G6sb _;hE"'Ã& S8G#X82PA*ʴB;:T&:mg,T:Dst1)8?#T c dS#k`E'Pl?j4nL)uH(M:ӵuZi cHUʣKө]dj1>p%h?jr<ᨤ0Bտb ' raB+'NL FrmIiM`M 3`tm娫u 1(9K;Fkj^OSӞ" zP% *i[=Z_dot;=F8ӊWz3=۪?Pw47w[$0a5S_G<;D"DC׫׿kD0=2ݢ;#oNй«LXgrBg)RrjMr줉?ZLnʩ**O)ELN]D3TwfCm7\bި^gT;H.eq9JyZoFa~]9n6VYSI:!G**$pR)X@c`+'h`l)) #6nӐ2}͗Rm|3U{sa أDU(er{WZ2X-y VcP#!xV'ҒH$q<(*3wۉ ict. C)LaZt$V۸EsϩI%:i35_>u`yO2{ erL?-kЮQ c:n^e]t٣y{'۝wgbiB0<ߛ0mw^N"pXr9@ yPHF=kCi9-MOn"dI ː*~!-#Dmph5<{n'O9:[LjJȘhMھ"B>C%p+d|\fwI߬\iy_ȞG'YkG"?YV u\0 =*z` ODt)"zCd {VSp%Abvlcc]uAɝE<ÉcOLꌻI}MgZKx$ӦY,$˸B&l2 Q` S7)gd0j)͔Y2?hhF1ԚSvAMLeo٥,mz Gu:dzOVtTfao4Wg 6)G,Ì6_a7fQp1O8ܱ u1f#.xFԂ8gb ^^fgB; r^֓—cDb5qBm[(Wնsg85f2\DmYHACN$JRMw062{{ZJ;rsUapzrk^ILžqbVC^ nzBrJU->yyUcW9ʏ+uẅ́- {iG#؏% {2Y WAۇKq]58R]7/YHiw#H*d3w&rp69ퟎ]KG/{ YvS;D'B: ,DjzhHUjxmr. Մ"[86S*8fw+1uj<_ pV>Bu9x:sXe2?L+wz7YݗA , )V0WuZ3yg{#M%˕s ]9IW_U j/Tc:kgu5Xw_w,N%o>@Yz2 *E|P8t?6~}L;cz:Pc h:q;1In/С0}]Xa&ҍa"wKG6qIt^2 98-[xo") R Ox#kJk7k/o傦*XubʄUQ1mFJ%++X0c$ :0š0z̆҆"k"u14^J4ޯX/qN\^jղw 5 lU u#;/x D̪=nj 5hPAΡHC;|uUCF4$0N;n j%}ju>#&eߝUŋ:{|h::x+ )p+qF\J$xJT.~J;v7}JF!B m6W=52FXͰT7|W]٫]- ht^ ̳:"nҹ(-{ j-ֵA˄͙HB)- CR&FU_REzvG >BY!-唕W0BSI$B4$7&&롕.9Tbv[QTTzÇQ?]b U"̪WA9=5A9Kkzeׯ=O5/~?~<5^۩F_|L݋?ȱL*!ؙ7#URډ_:V"{.k> mׯN]&h\+ H\O̒"_߸J7+xCI}}ڏb9T!RWy&, ~#]5\Z5⮋OM5D pō9KʼnW,8mm9\nlϖs\atrFpa'խ_)"?0;wNڞ:x/D zPDZ3?(@ݏ tCea.;7YR+m8U|;6-e%8*vRU`l{~FM5Ņ䏿ׂ:8cw#(^I4{7\:bjlB}>.D沫+u\ U{w~#M'zhT](wAQ2ZY{[_sA[yy)CITWY,aʀvQ$Y&T+U{~7 Y+̮n+@@{ՍFNZJ }7/]2[Ō1*jouB5msZVim"U1=ḭΖ-})rT0/.:o"Y{#VJG!/1=Y.#. #ΜHg+r=9CBƠ5fIc7>5悦*X:I4mWrog)FnM%2#)D;TxaE>>u2ytA+s^}~U-TW^>(MJr9_v,5tg]w{k r+ 8sD*E2O /fq6>L ye%继xĽuUiRZs`ʏUeAo~[lujBwd ӆZ=<;hE"D `M8pմ7dcuйwvqj z ǧO'S7}-T18q@DEqV8T6y[BJm0ە[OE+ET"'gԐF8 9#3ԼRYr<캠f*q]%)C,2@/H,x[0|I DP^iI lSF`du.R7N?{.]#NW2},?20O n5!HD4 5$. )4ޢm<@^7S(I!UĦA10DT/~wjk)1岜fHmKo#Fe F,?;` ;zuhOcBE*d+ŔPoM[u);uЮA%޷s#O':;. ULqYZL ^Si JjiV`2m 3b!IO*d*̟f%B@;} 00`u!X4%:F{y{d?vB"utUuI>E/EӯVQeXҌ]Uﭴ1όL$. bEO "YJJ92﷿)&5{8? ?v\,zlEgB|wIW.oIWﳦip[g\L+` qBƠQ7fydFs)t5}>gC@XjNP,F~Ԋfm@75<{wN\TY{#%;WT >p)( þ$9RMT`uX5]F.]=G :{yp*}8+K\=E) 2-ӴDKTV1,%W?Yr*U*6-lA/ڒh9m a϶u&~pS`dDt;C®tj7,YD_iʒZ6Pφ$rxˋN0d(O A\d8R(w%Dju9osݘl̒qUK*мD|' N gbB snϻW&Lwuu1C*uL1}e.h an 9)@œMP w%%3y^fp[y~4/=<j^f`ӻ)Qׯ,Zq+B\S)}6k3t?5/(b{=˵"X5!]k t1 0x8Ӭ?0Ejqn}N-avMnød$ M6 IRm[S1sU1cp]nnZqCh2]l܊fgN`yֵ;GM-3#YFN/e 0쵌hsHO' ǘz) SuQ[q?E^eH?70Qo t2K&K q՛^ϽѰhL*؇р~+SD@?yq_T\q-k s8N6QB4٬$62<]7}P2ժ4ʧᅌ.`Ujv|XJ3NoCsL pu7/oIYE(/&Yy&އ,Fj=(To5/aAՀC1(lR.J 455L7ceʕ_ Q<^$7E5S$Q背Av⃌$Z~cth4~T@D^,Xf$cFs,4iY4(<4:P,#N!UbóNgNe% N~{f2r|qsbhACƘ (ηl,zv>Yf>oH񭟆T( m'ܰr{ AhPc1[nT2$z^xJ#s}Z:bQ *vsM/Ahi*- Az!ek֏RZamr(;!k:"68hzs.ptAakᅮp Fgd F*C$rv2d?Yo<MfzHoV>嗏xvDwx`,QrOz?'Y1nBo{].$4(f~/vH77bP>aV ^Ňbw-R}Zl‡=(xX Jʘ"^R*`WRӁ ;x7L$)Â!3%`۔,z 19@k6|O(iP5hPHڅrbZ X HZnOkV7H~DaJ+"I $G| `Iζ׏/;~fk(nggpA͏rv[دwV*04Y9zZo_cU[^7*,9yS<9gs8~5?eKsi3"* ;iP%J dIbz"?_e pfNy<G ɝbnFИwPDih8 *c/ >0Mj$qqQ! 1Cz4Us]+~KM}>LJV2}`+`IkJ) s q9Js/#s ZQ\oT*ZeL8`Dcp%bTv˳1nUO[5-Χd8F0\3F݉ۦKfra UO}2U`A:iOp!1?Fp¼,8Z-dPǬ*T%km>}0N~L%dO(,FQa0"Et?ĕjڼO矷U g`-̃q=\3'E/<,Vybm|klٴȳNoפֵ<Ɨm~}NK$pN!\"Pso5_32M%zr. Y{ Xx:"B?`rIlKW/ A&O=!c E.@btW,akuc] Lhr?N#6\x\-wABu%]% jdڽAJ̧ :3w}1 ,>ۋrkγ_ F_\?وcJ }z ~܃P V`4l! (hM4^GP*cjZ`@Ež+qG`pDh z`#$wl&7k"UkB_=o #ݺ:#U|PbϜk^=v"u\_ g.X_w{hȕT.V?Tr/|Ntd'BW 4 \"I,W тQi6*)yEFdCU-A;d IXL?gϡL{WǖdU׻c'>X.Mr)qv)27I5s,IY(' yB=D97ׯ-"YB T*Lט!])2Ku'$W%l]YoI+^vi}ì40]l<6(QLҒF(xIY*R6`Q3ˈ;+)%uFUju?GdTKӞ$ԏF,KAIx8d@dLNHv!XeL#(Rʵay`?nE\Z[.Dt]q2~Rx-4u5 w^>ǝYnKIgZ5o&5&qIDd ࣀs(, i?X;UJ9)!䘗ճ^mTkEH`8-A%`#;`$\:Wh $r4'M0ԸjF9"Rm1#<3$zKD@\Z E> .24YOn'5q-o IHPR*h)KC[h#.oƣKI$0 3@zx;H*Nd;ಓn!&xpz'Ys'T=3 ?Zp/3vq㭤l-}5= ڤDQ2dHcmFYpnwWOc:"`PBMuW ' [u\i^"'pW| ;V9.JQZ'$GkBc~BLhroU_Ԟb! f4Qid1VCu[^~ q&u30/5N2yUK>DS``.d,hKƕpFܗ ?h"'T+" h =I>cј5j䍸d  FZb)0@-"ޑ| 4V>uPB ̍] h&WR9^V9~pq/n(]8>_*ƛV .)t|B].wn :?cU(A0 o'lqA9壆?{/B&$oM5z\Bh@=~PPzy=;\JヤD滠z2:ocgFNvO&vRePQp .3]7"fv{VVRN3J+jnq;t+Cim tve,٧CG2F Fo$oQ$LJg@hJy>5fC QI6~5x.% 1Tv#erAJ YȰ!>7XQ ⨱S8^'3nBrb<4Z%5&Z3Z+!`h]M]4(s.saAt2.<+S#TFab%Vlb3!/fPND@qxwv^d+cÔ8-r\qC-[{ tSvp >Bw׽^[DBC6 dt>!y=Nh%S7YSC~ʜ:]9wa8TG1;ŗfap рb*&;`~ Ň?|y\޽Dx?bӻr~PͭϿsq\M[ʮ.f~n/o墝 k=&6Hf|?Mqq>R ? s>; 4FN4{8rytEfXc03=sha6j/xC')IBa#8%#CLTLyiY쟃, ccފ +GjN\L'YWsch[ڱ~vfƴ\ T]?ޢ]wϩ*$*ȿjai;N6*hFJx9#d2vL4W[qC%}RZIn%oؖ?I4tpCFlz,ݐѵɬoM+A@ڞ{DajA|zsKnEE)Xu7iLja!%y#^-0N), hy\O*t2I%Ggվ&m͝쩚a<]ˋ)Y°`Mf x;4*FW#;ͯ8&s&3$?N/xfK?HLY a-H5pD1߱-Sl3Q)dlml3|#yjj !.[hyxXp>KCEBX@R,80 ?ׂC|J!___W^Ckv(FdwQaՃ TGeyP@g!$UGa [vU_rPf@e P5Չv$xcXg91Bl?{޼$XW+͙5w?6vw~x`ҵOww]ň]^=UYqP3CTߨٚx E<=8m#i։>{YUg?@PREŦz 3TC} Lp[ NEeFARˣw -߇7I$AiϿ*~]lw چOTnJ!xaڡ;k!ߎ7 5]2}ڵDFJVcpϹE iy2H/sȜRGsUU\T!Λԉ41 !:3:` ~U|M4#0>"ugŭ?@1J&2f~c1 ~2D KEm,$b4%DAia:Z`{QD"?2G= 6Hù:v]~cʯ faԼlNs\<NИ7:`30u8] 漫,WJxTtc[Aw޴xΧ>73VlW7yOJw>.ފMCmcjp 2=X\v<*o9F֗P.F_Oc$ꭋ*nZi}m19Y}Uy .a)'RW8 4U /k _&(XOOYBc#jn`p<[]1MIwoMߙEWf}z=$ "70c; PzTHsk77vX}UyDȬB;|ޥuW }.7|Nb*PtA#]B6 8K$N.:I he{l7:ߨW@)Yw- Îi !; Š=Vl²!tGYhf7^ 0*y4WY*"n 7f)Ývp8-|gz6߱$EY%gb܉\a&y2 S"ʺd+,bF ?QPTdx/-m;S`}xup3a==/{_rU__ ScF sog7)U!̏4;KsŦ-!Z~vV?;)-f!-kGwVDvqu`=>ob_hu'KߛP!;Ee>n:=֏ܣH >?NTu,ff+p_g<ނ_v.9w><3ى37xkt$W2Jo-6𜘮*BDg%)}3Rab/PwϹ+9ƶ!4hw]L}coo,(ZO6R wVbŭ/d}C^1^n HrW}]Q}YgQ~ D~6\vPmO^Rc~oLzSRg}SjKKvUﵗ2veKW,d=弛[{ ӽw*1ۀN0+Lx-t=c}%APZMi? *X`e&3̕ (:btÏk/8[L)c/oW?>xdz'*_?>뵔:k +ϟ/7!FDu]:_]EyV_Ovf'6|$H#q$ <() -}@B7F[ͪh<$ЋA+82208VP>PpH~9&HA沀Sٻn$WTz٧+h4vv2UI%35;[*|ڊ%[8=(R(]Rbh4iN\@,KlFZu6a^/iӎ4m@)A8rw?8i"C;+ƣb.)Z,&䵆٭o‱`К>\U<&H qtnnn!i5wKfI>i8x~]Sڋa,5G|[ĽJXHXEWST$>F-q胆0J:Aq)Лcɛ7̺&pѡӂl=96`k fTЍ8-8ud[cWJS A` 19Q| 2xL85mm^k^Ê?v !WJ:L^?5}Jcf0Ј-Gާ;d6aw%Er`xlJ|~af!V0[h A0DT*e[Զg! 2bhh|dhWYc59F,\qV{|XoN9&u0b A ')J%!yU P0c !H"}ڦMVE$Po4mI`S5_c7qQ⦢"[Q#ˬ2._A`ϊBJx :=1MyL>[vϩn*6lyO$ck`g1+3 x[uק#cqΎ̃ސET0ΚNgG#웗N8O>'es;h.dg=` :D;51bs=[|gXv>=IP%ScWԄ th'}H)^6sbI cO{}uSA;LczM6/-M^`#'d7!{,pG᣸)bd\ 6~Dğ^&)e>z<ϙN"=)J` R!m nd/*&ۜ_K.eم\#Lw%hWgίcMÄ_k>6s!;??@>?NbXgµ?ٱ2;WJz7/խi-5C_G&!]XL-Y՟a@S=~xzu$> \ziRA^O! %?}kI.kc}b~?ۥ?vM8oDn/T Z^N[EyMg}5e_hh&dm'x t1+Kj_a6΁m{kOԊVf )V4 m ij@>lM5 zr~i s\3rtp:;[47[/ec.vuʾyذd v5iV35:!ዔ)DČlrd)B H} LgRMZ\lTV 8x͌/`eEMLppNP]dNDpy3.h~eZ^+V,lyC.9IRTjfaIZh0bJ#!͛6x䂮\͇4z|Aa>h 3Qn#-.*k}ENlj2ZfJY\j4ΤZ 1MDArn|1Z#tˁ'_:߿E'#Aƫ C b*ʼւB"Chܘ1){EbŠakY`-s.ZpB7cm}oae74C|8aj2tBS>ak7B:E)Y 6 Nʃ^FPOt(xv qe&/Z\zk)ڗXC, H:SU>X|`7H8[K_;ԆO 0(y׃ =拁BE7vk[XxLy3嵧eƦ8jTp-|ZcuZ%`immS< j̇`#Hu ֒1ƗT=Pl\0)Z'W„ 6D+(BPM&3dGM<@*ejs3Ml+[0uBCܕܟJ;'t6[qچ`؝^Kr=ivu>,j-F}-Wއ ?;>?{ Fpvv|{? қ~쏯܅?;gxgiO5]1Im̠ې>?ލ|Z4ߺ<'ܻݟomƐ}vj7s_>3q93O65 !sSᄌNSn%fOʋ,s 8sN)Ҿ9.{[ ㏆E){%/P?q+}as}aZ+n Fon5j Dwj%(5r^gC4?B.Bh$ђF{JR4OuϕeIe.y!OW2t*(1?LxN+)F+gd+OӞ³S5l-״tS\^츟WSWYMU޶X߄a:DMU̜#B?2lY(ܱ}3LzJkLj =iewTyv>p M;\T(вc%,vv[a߬sNMwhw&h'1&b<ўTwv> co'hw̹=i,h՞Y˳ݲ} nڐޡ@;1b6ò3Ɛ! Mn%hkpUF3r3a>#gGrXk<@>x7+"|gLJ=OeRIijprŹx^'LO3An@*R r=<}_a߂@ijI/nOט;3E 0I۫}h6,zn/{3)FJ󩅔AbU%c3a$ -8BoUO 5ߖgK`;2xClRTrkU ԤMR:dpMXMX,@w11+j"Vit5lazO%PLϗԢ#ڸ.1?ׯF̾W_B@D'Wm=\o]f_k^ُ~(ypPV(_تrדujlzg0h3:^a]c1kZc>Jdc罷çGilUi<: qrgc967˺7X|c z2| ˘ԲZ+ +.;`A>H^`NUK71 `^hkp&KI[4MN bz&c6-Xo5B6D+"lz"ۍ߹i/x^3flh߽[Ӿ/:LY2ubF;cx+l@B /S+d0}\#-%Lbz=U&TMqk$nNS~Xvb6o7u9&a"ޛӣmUrR֛pنTnUǵ1e \~6G<֕S ]]sS=+W;%ZQnLOI m; U$QKec|է׿:_]_=Kd:}ܒGT3Qx=o{$F]`9`Z7(/U9ܥ ח2ɉ]"D}OZ)r#-9lcF=tlC<0a{%Glj7}{粭7KJv(vw2e$NI/N-QNEyO u</v7,JmLI/W?XKAc}^>b"RYt+ ZcV4˻gN‘!_Y@IQq~:il1ɥvPBs7o ?=-ڣ//0 E?>}ܖY`0J%bas.p!d>x,p[ e*G<|@l D>ĸ}O_e<=1&D=[;\a%JULLPoC?yh^'q} 9T YX1Lom-`q%p31sN@9OۢuD"'Vk=ZчDhEEb!g/ty/F!k~v=(o׎*3I*XBUJNNuGL9H:s]`GGO$%D}[h,΁]m^}·WmaXdch/gkWntybė^xEcE{?toY^C+Wk>v??_}˧hQ[gY_d|~׾z~j(WkkϿ}@m/CWϽeG Oid~BA y#pǖ;scbab SUL8 Ͽ~f59)ǯEUR uWsy5_|y<_3E~,g|ՁJei{`>S-g6=%x:b) b3I.3{$n sZӱZ7/oY9őw3.F.{]ex4|-$C5O,ɔN֪D!ȓ 1ŀ;|R 3~zț4}¢wkgț4a00s^(^%h>A-Dݣv`>%h6/FE眉a)9dmw;kV *͎Ѷ;r&3 @۽( Se}X-Bh)$#)g<|IUSԋUZ[ U%d%P]$O Ts' q- !ʻ/֮ IU !jn>gUYS i_zAdWncՊaNxM$k*G||?ꇷ@}{ĵ lhl5@Qg؆F|`zU鲋hl@UQK%sD(%'x 9VTp5Upu唓Ub9D={E흣 g9J?Gw7OH'HDGFA73ax<4}@0Y30)B1)(PRLNU]%+FN8\+so:'VPRpwWn 09H&?8BOnQjn/V gþ%:fw6w*&Mi۞[&qu__|OA֘#QVjj-RPTGғvبNp Ts\Vr/ݧ$O++ɓJJ#e%wE̔2(S@gP|ɠvTZ+AnߓF%;HeNC[ :Ċ]>N#%wلN0$Jvݫ ōHTCs5Jk)lCYWJ@Z\GI.U3}^Av#{g%T0*E9 fIBP;Qebzg!$~1:b7NGzd墼V# K}+Jqlڅ>s˘^WuHM +w]GҔQh.GQ'T]B^XNv8Cws|a"q- tөn8 `S!# Rgz+8RFl­c>'U->fRYL,.HO'PMp*d5i~hаY[0DVe|Tzill̈́qX75q/=6BQgU[YEύrU׶9+\P);rU&82#tJ]OUe]'Nm8VłI*U4z{@Jp/\Suk%7uu PRI@b[OKU󕖽t%Tf^^N፩"X 7kG )uLdQ7EJr-$4U]W#]eu{ErxʪJ(_䝭k< IRy!O*")U&(e}1C`u r)H/qӈ>qХ$&0 "&oy3\X &D- 2Jpd(zN1.zv+8^8yyD"'tnJdȔ/(JzҫT.;jI]߉[j zrj}i"\Z׭,6{nӓ zlF."QU!TM&],׼2!.JAR S>gWK!+l8J$QK`lϾbEgv8!OcW_J>q=a,jKe[c'_ՔLrH]'0IltWU򒲯}֪W=fVMS5쫍OXXBfĉDCFׅ AC%&dED+#VO6ۆ nLj%0ޏ1I^ϪDWܢw2jSյvUlE/:Sv¼ I? AL-b! DyFsbo%lȥ(>WC&>kPcͫv]br9Ht%FlڿU Z 2@NcQ{5Bخ)ӜTQZ\QUl BGE[xw:a"d׭&%G03>(`VrWaGgۮQ%Ϧ~lRlUqj};53[W9H |8 2bIuy?C`|XbV]PY8M^KzYBɇ5`Җ:!: bմBT*Lqd[Z{ V9Lwݺެ;cl6'\/,&4N:R~ft{!<)C-.Br^Vhc5E_^r>z*[ߺ8K5/qy^(&Q -NpI0j=}u s& frZ6TR@#iju6b ܋B`(A5v:M/wEVU2,?]FMɃj#R.bLD]T+(?I3uoۏQU> "ItE$ѩmyHL `r2@5.xaZ3/986x*[ak %Sjk%`Fy{ޢ[BwS ԠbiZzD,$r~~qɓ X;AZE\:t b1 x/'z1oM>9;6r]2xۻ%$s)[>cɢ6nM+^n2. M *$P{n62Df x07R9v{)KR;u_J9|Cm-$oḚxJ@!iTywWYWKeg8Z%4ɱ=i}di)!-ɷ}-PR|nAMgk 9oܘY*5]^'89Q*W'g;y;~QY׽fA0YB"mnd?UB*+ do,s` ,#Kv2'i+j$Ӌ\j|vա!ֳԆݦ^M0Wrcm%߿o7+n$QrF?_csoz, #ޕ-d<].8O:~*gJ-l{qm{}mor4DHf ,9{u黈eQvk;sצ]u Y--~]gwYOV[/c=( Ń~ Ddrv_b%TR%nLB6owv;K_%L/_)պ+@V/1%6>.CX'z̚􆌌g#zkUOdȁfzo+zoLR &Qv59I0\Zc='H O󩇿 =Oݴ4z/ɴ0rbŸ zJDdZLև郶޶}]bZ)kҾx6ii?^9M춶`\\%a5MeZ-?]~|d e皖7rۡ;ziجޟd>_?|Vv5s.0r}gڅ?C~,&׳ZDԘ?|w"JxŇB?oCc %q1g:E~olf^?1VnW.N?;x-&46qr% ](?^哋($S՗&cQBYw Ի!U̘-zGqp<[ ](_P]w]8({K+O(z"S!֒lu b_-1~/yoNaG=n+0f>;C hs 4hˈq3/)JrbKz"6cR(V|_-ײgZ"RG%OjW;o,6~eo}҂mmJ|8OeN̼B3ZEzTސ·ɫP)|;gtjA}W[?$"`{3#42"P)z^/8_8mYc@C~Q=xg㷁و#Ut@F+{5_kk) o~cb%g6)"TSbBsɱ,xȳ% -.xOggjP]< Dwެyt2{ҧSS9V?Yp:ա(!ס:H6ߕ HIPc"v+7b?L-@AnE"YvMXF7X=4PY`DŽ {B؝4Li:bt%.1Ge*F8pE jкdӠ-U[ J6s6^3[kZ7T02ڐCxEhVǒ 9iVlA(D'}ioUG@br':;=]y1$[KW]{;F&@*fT03+{`Bt`,!e`X-ځg;l"akj0o-Q R_(}Jj~`MʄM/.:hA[ 9Ƒ& 1jlJP}FRɭ%c> Tc4cHHt`[5W[̗<3C3\rXoc>[@GRHZu }rg#)Nnz5a8R_@5?#yk|E ޝOQ)rJvlhȩka+@b!"ֹRT*mQbL+8P=unLPyRK*#Dl,?[:Eʱ@LM)O2;8@) jM bfƯOҌQ`[3ضj UB"ID ;QP#Ȋ5WkѕW*W.iXb:[,!GYȌ LQF٣aDjFj^zrV.֐ىˋ܋A]%Oik4V3C/h)*0۝_ ^X0T cˏ[ fDz*e&V CEA)F6frjX<`׭|[Tt d`'[M 3%qd{v< SH u ,JhyٗDBL![n8|)zbb1Y`|XbB9zt.a 2P\ A0>sĄ(? J|U1{8D m]x; (z}|ɳ7F;0+Uv Ǟv]kF1 p-Z"ql5g+1I-j3l\ВjbSA@Un,TB0 .ʘBRȵ~LGRRe*A)7%-f0ZOzv%V8h s3r2Ki`ί.G'6^5kg5/ ʋ,ZCW[_0gn s1eLɷ_xN 3F2}xח=S-Io3eaɶh۔2rׄՇP@y/Q"32J%QkMhn3[iPnZ̽{y# DYP߯ȣ7FQ5ixU<6u4Ndg}5D*'t FJ&Sb 'WM.nHyњ^C+I'_NV<ήdzիz5l`uY5k$Z+|ھ؁Y0`-QiQDKXXPEQ6EEC; veakϩFMM䛁6ʜ!Rs:>PFMXޠ:tykgF,#:1ۧUZ|eۤz۪.7(BBJâ*vZClXb%Z}.de0Ҕta~$ |l#~xa ޼ӫ% :AH5Kċ-oURF_gce/_êsmr"ƀ"DYf9 io5n<j)70`'i`Ew'{?qB" ,p{Sf6*/[3?D}Ny`~d~8OOpgLZGr E#gh="'y=:'K}4n;Tln]M~@}WpQTHHE"\)f PQ>l;,];"FSW*D+񏠃I5(/ԐSM)Ix?E-Q 3$3%e&Ϥ rpjl;t_,z" g UȖZow<V_+-2$ gzj,9-\>`A2NC4l֘Z@ SX!V*v]IZfy'P Hƻp ޒ4`u[Aw;l!Z2ǘRNUQA6D*s\O^ɱWف>nW֠km36&=хE+-q##2) q]?{@Aݓcnbu<,G< ўN>wǾq䖨[ Uȱ`ιLP39nvǎUZH5#'Y49ɲ&U:["%Kl1,#ʲ %"·*ShT '^A6"|g@)gᣉ#G d.,3M YKy!ʶ GʳO"!b$S,e=b1gon8휇8L&Gfu;٨.lP☞'!G٘SZ[.-FvxgG(3G;=(3}G{zM,"u$ЁaHC Y4K.p|c;\RvӍ>I76&4iTʼ(\ "yM(::-+3xzG{Z`{s[Ye|Խ~45}wKn SqڒFݔF kvot;RGA?A2R+4x_︴0XڼlCkn9٘ј)gq5#kNI1vXsW֜׈He=,[lDN MN{ u a;Q莚Zn㕅1W lv´ȲOF5_qi/!mn˽qI-82G_!Q0^.-)\yviه-ABſ -槞˺57 _Ə̉?<+kpp|f=Ի˟*IYH* 67кwb C HMv鉻M <%󌵃MZo6uɇ? +mbF|!"cw4&aDډw4~v;w4ONc!N zxƄvnP Xw}xGBvV\ WZ9 wv{4*<6M;HAGG+#G6sksiS?Mj,q搧Ii5޹,΃ADZJS,k.|nV"?.?g7^0CS}#>t'H mo)ݷQYkͺgR[ k;7ϏcA'jҪ?5G|j%=FQ6oklb׬cӆg՜)o﫞|1:㍱I,iG|G;P\5l+_n}_םzثvh믤-nM?F.=dae&Ah52mtp~v_U$WH vxХYP U$~oäDӂ&LlYjG׆.6G#|+GClm/vKPΐ,c/D6_i,hz_Gwnk>)-mwI3t#<ːyDEo#G5Y9Zm5e1ˋ%ArY"6%v0S\3~9i)8)7Gl4XQf >r]-Cث}}VC, OclɌE;$&TGgpVo6R(o$i@`x~iU%?Ǭ4&)1I)Y+Kcv,UV 4u:).ޠ!ΒU"͊Ƹ,ڡA,'bQcߚ}`U_9?_ѠǯK r9*. \#yOPG :ܲd['9cCb-8 r@@p2(&2OѪ|TqNy/gO Ǧ lXfO;.)9nYb8}n3~ݧ<x__/6NrwӋ[6DQ{l$JM;&r*E#8 h3!S9XtpWsgH:>[vI]]sv38!*eu3><ǧ'a!"s!8֢je ܋Hazz'AQ 9Z$h$9BܲgINc(,CL@k*l [ C æjT3Fw5nP ar$G"[I!V03> ~ N*#<33Y"O& 4U0{7~I+n0EZIIJ<m&_-:ͥqƦ~BgC;`<_ypw^?$/#ax+pZZ&W4zsB{ y`6Ϳ6p{&՗>1Rc)pOѴT$ hqVX0kX{}Nh8 wėohuWŸ>FTApM,ʚ~F{M&k\ק4n~4-< 2Mf -j+Ze5"wgZV:2N?L/܎nnn~]ܰԳ<]nJc wvZ5f1[[sHYMdthm[1:q3FEMN7^F+(憤,x853K=Gv@$Ҿ9<["_u w\mh JrY]_x{;ͅN}]_ޒY,cQ}QI$`})uE)vjU".qZ9w tP~j}˗{sA ך|{ @dTQLX5>@jDV7ru#Qiܰvw.#&Ӽ0*ի;_=]ţ\|{Kȱ+TmԽlv{63vy(PJMǭ{BDI/ʜ>`q~wW|LVphHj-jX]M:0/ lI0˽gTt> O돃vh1Q ykeg.996?{!ѳF7дÏߏqa rCɎ0C8D+a< (\DƜ4p $j .BhE= +fN&e4?7t+K$.$?Ps{ƙ{ijԴ4OwMBlލJx h%D nFeBpi9F睰,AA18>6`.\oN=MF6tg)4DK4K*'+_%&qWDݎM.vowy/ƣ "q>pNJKM8̪7hW逐lҩo=۲Yn򼞅Y؊843wI M±$8 (!?)Yyc5t](AQR=' HY%&+R/9pQBKQV dz~N [QI $ac0 m.%DYWAH`- v ‘N+~VK_uu[+CD! V蟳DE0yAq\2$&­aO+Ѻ;Qṽ,35ǩ2Aqu%lw 򒜯ñZpp~+ٓmAзYQJ:P^ۮ g)u|JUO@iEWvhNSupD1dѫ; J-z~Fr|-e5k7PBd P. @fh!4ufQP{lj.MJrwx]sYnp ]ߖidPD\Y(3痟r83Y]Nۛ|xPMB.J7wҺUIO*yI,RJbRX]k)?1(IDfT}>8e`m6(H]oǒW~طP}!pF/T3<)̐lX$5Uwu]]U Yb kCj)\1ʃqi3a'j$R.so̓ $ëmiU*BQF-1E`~f#q֑JsܥotS(r`R 5JfZY=֢@Ru,7JS0 Lc'\[g:BS|RZl Uq"WpU7qx}q,;*¯ms)\d=|@77z1D[]ryAzŬ6}t.)t함5 AoqV-V36W ?; о->[ZʻbZ>ݵ_'s{r0kv@ެ˕DqL>(F~y,0RIM:/(lSN}y׍kK)@RvX L-!S>rՎT;B KIC%TN֌&RL ht_([+më#z~W`jƀmf 8EouįbXG*_MGYL\xLHN$R,C;Gqceq<#5od!{X,/϶*1g=́%~B4g&,8~hQJo[?a7g 2 ]Y10$jami8jx;ixwNK1,,ɰ&z<5 .!Sjxn%5!N0m8Qx!p90nOCi0K't]jP`isΊ}]*'4O:kN*g1-f<ְ #8^l 2 CA#o)3nklnpZÂdEK7`5h^x6 qrun DXDw<==]ˀpހ-/vFyM筆ϥj2MüG\d>f_M]~1CK~)'P;?e?Q#,/-X$U #2;@~&7[RW9ڧ,lC^8^ Nx߱;;o Q.yn?ʷuO» lLe訞/$e:Ҍp2XyS:%T1wFvM P ZlhӄOBğ@GHA_+\J1dxQrz϶ Rל8NU)~2W9LUSa阏3okWekyΝnaLXѭ0->_KLd,|9C%,wÞqozW-P=V/b=igQon-@ިw,qnr'dfX/gSH(e0!)n$;|h_9!k#`xzvM9NӭZa;0ǩ WY츄a~67S ~峲ճʳⳮⳮⳮ#l{6晒WiYtZ ƽ~~k' [2n[~,#dQ{7\F] ylVfky"gTLZ:ӊ3}5EQǙT:\iϳ6'OSSsWq|C Ig5&A#z줵T|*19Cq!FxdqxO5Vq  DkZ'VYr ,xXue~m-8PY-Uofa龊jŚK{䱓hp,o4zr#JP1P9ԌuL fdT%Q;6ȶI/Q 2v~QPH"yD"֤|3׿+xc3]~^|/P}]m0<דanƳpQ\{g{8W!{=0t),.Y\59K= ZxOs5rmP |8pC˳E_YWB.z(,_ |e4xH+7IWq|OJq}k.3za1;=ūWs?{_^qLY]OO`Vxx<;GIBS%k){<ǟ::3G]~H5借+IS>&48'I/^π"s.~ `͛{>H.2-ӎ7ӗRFϹȯ|4r7Vc? Jߜ_Wln̋q!ͧ`ɻzqc] Ԋ ZXqVFaoIR>KV<ݩe*t +Eojl`7'QC 6u,ޥhKsNmA4 C>fI};IUy{(5n}o>xN5XP&; }̏w8AQ0ta-qWʥ0nZn|c1'SȰ5(AcHdSckQr"w'C5|`WH?lW$f(zC'^-Cȣ}bEquw;j B̸K4CYmݸ/cL$N{~zdyͲJɧBv<$_ |<<ߎv.R}#v]KuR}u>d&QMÂSY.4U!FcUmGnʒaixf(ĶJo(s9\xus孔?O9#%ROa"|R0FFҙ˃8c+Tq&={#WϴF:oy#!uHtNodSO =M w#GրmlA b'GRX4ܑʔ ]S_mXe;ZG\֬sVvy. q@C7v kh%jUUܰпV]f/@Syү6JMwhY<]w.;Ww> '^1N֞`ᅎ(ύ Kͅ)0G N0Lm!9ͦ'Ie~x޾:.1@.Aj$|fo:˼Mh]q\5]ٗkR:!:ǘRTRj#JF5%4"T!>Yk 75PR[C㗳q.?\6ecy#<5ZD> cS2^iLˑ .RRRRQK9 ۀxQ Mp6("OBr$ *iO̼$$b朲(5Ȑ ^,Zz'Q^2>?kBA\9.PPß 3R@]dCOюX2`8A 6URv|" qcյQTyk}eK62"}$yUO#me/1Q&2$"z /hpEqkAmm@:`$gn]S-#EaqYF~XJXVSq tHf4w+֭2rɸ ~g JOb3>bNu}c봘Vvػqx 8Q/mɩWeu[kK~rqS] k3rNBћy2/NR|@d~z-p`2͗ū橉+|)~)&xX#> p)h'_}w~OpƕQ(eZe=0nSҼp*hPh_hCLa c8ui:_ZGܩBH6 g:07{mv[Likocז}}`Wnwؘ!«mP7Lmop0`3UO}}\E:fL*l.Ѩ{'[}W㇁3+oN3EM$ʕt_w/*-18,4At p%U{QoGUSMMD" ܉,o Bv`C3LDp%ZA(!ː;\ys9|`Y sgU$qڨ{v;kNNkNo!'Dޓ"o"B]h}EB]hU_ؾJev VI:O #أ"o%]l?ܯd-/k%6>H NWzy{*L1Ӳv>Y?]K!V`&* Fx,HEb ɠ)Fl"*Fo , 虺^?u1rsXkfmC(D!jY"~sd\q. 4J/)H"{a(E{YYסBMp,uHw.`Q5.Vh0O f -) V`6HeBBM{nCGD!p9ɕ" m5OJyIj,u AQ“S.)lR0'T 2&*ySBz( ;"`\)cz/ ,K˳ny S:1@Bl4M -%7v%> MOmߚe9TДIkw&l~',!;0,1C,i;6-4nW*d. 57 `(y PHZgXHv-9h#; Ͱ(wqFovlT GEtjĕfTa;̙NжLY #X#G| 09؎p :U5NgU #OXṡbV1"Ku#`rK_#$ #C6aJ{y]plVѢ,L`)K}iUTm1M9se\l72.u׮\@~[zx]NOS4rz YчVCDex!ۣao.zW(XIdXrŽ{\;Yl/د)/\No.?M,sz` n*Ҭd `[ /ƁÄhSd=m|?pV2GwA}hAvc3&4}3pI9']Yq3%GG$1.xVu4t80,e~a33irQ&;aBAj#1nYp Ji<\jq3LFLO@"n/i뙲HzM:]@BÅ҆!rz)3gZEU0&ѻɛz4$ZJ/X(RG:bAx1Q^SWӎ`i>0R0]閫;n4Uݤ?;K FhlO28[>A6]AP_@xK%O(6Xo(Iɒ6tF<m ^aS :'Z%oa>+Ni Rk0vW|%OiZ-[0"ՠb| Ԇ$2"h P0A̷v=DbuoKs]^(9#87dc5d1qD"{ zۃ_f HAUw}&C?rR^:r9rWqX oV TižM_YVjr6ֽ/_iMSQsTDce18efv,Jo CB zƈ"AAz _gFQzgl@w^}lrȄP-08.Vsɱ'Iw:2x`)C c-"w DaC4*2*Ԗ{":zZj"f@ ιU ӹ5L-jZWe`?s}"ćYYC',3T0~`c:Ņgj۶ѷU_28\i~yɀg>IN,$٢$ d\bX"˭W%h3sz, yڝiĻS86)_zwnWVQ79hVPzNj-=sauqN|ι0bf\P[)),%:K1ΘY_0DawpA ҳe<+"ŜQN{Ng[ "VMmL#y4V-VF|#:^ ZuDo9bjq"oCcIR_P֌U-oɱĪkd%OvʪH5sX!}TjD!ߥ _=p.x CTfq(|L#|`SBPb۪wJ5Bd_rD AUK\-wm5j@a[wسxʪfB=]1U®5|H dy}YxCwݾW.%DuULbn_O~j({`Xs_Qp: O EhyHnN!~Mͷ K m~7Dz% #4*Z P.+O M{X V5(,O r$ 7$F|eМ/p>/t2>5h8Aij]kEU4:Cgj=fi:E[F ҝgf ;hr1k|[_>h&ˤr~Xd{m' -%>tE8ڋo{Ǧ _^Ŧ _Ǻm?JĚhߎЊR>' {'x h<7ŇSQ(|(9OR ̺Ty0Ru>|xRrv$ 7Ѻo׫C/|GL1']lk8oft2/cطj~rј+;ɸٰ{뮘1g̒pnP.=0^}-Gf^|Ԕ@hVRc(v͐}o .3-(Ah4Fl(IPn)5׭c#qP"'U(cV*[^~ӛ ʅhOGe2YMSȕveS;SU[<<7@S`2ҌJL:W`WS\|Mա AEvFU yA~H}7% R aL[ Å`ߧ(9؊2(=Wi &5AVXހ_%ճ׏ l_ӒZ\ߝAL|Z|g7˥?:}R #n&LJ4 +&:"óOJ0!^e@d'}": <犟к qV|sS!+1zsX#,I"z4IAB`sTM 79+-a?nU#4qS'RB<3848t58Ф_4#6E ]N\T=inU=yn0 ~ѾXz2v6Za'|8[p+/Q;zhv {;?ͺ^+ovkdv-_LMn_ͲCn QgKN%DT.=r[Jv%NS5Td#v#JX<꫒abw8UsS>*>{%8#- +df.ɭY*\LSGzTt9mujnc ;'vgYT9 4I7ϩ-wz[)z6aSxIM@,xΘm=Œ q8rk:*Osmk n❓,R94=YIP_>;|W]4Cie0AS6AG ,/~.N!T ;vNG0&e8"KY%;897~r>a} ιSIo{waUӿݯ6^N}߱oer\()q&3xX0e1-44XDz(#ZhfVW]D|WzoRj+Ӆ5i*0sg*Ye"PmN(%}ЩuJ_s~|HcWQ/|&>ݿFα"G7VQp%rOBt$ix{ss?.?q3E_$(tEz ˤ;> ..#c@hzB 5tibB4&(r.mt 8~be ovfnt;67v{jⓙf/6~tƃHn\Nl^AUۈc=y7`z<6ߧ׭Cp}!QAxtuJU@Gewnx[NƸ!B)j^YоVgX#=T< kXiXDAt+'Ri$^r|rodgS~rx&|B(&H /ދ Tם[>)-V<)Nmlu1#R z>F>hH-y_|RR?WOɇe=ٟ߃qM}ӕڍ~.? gMl"N]N M v*q q]8ݞsWryxϧۗ/^0vIFTNrD\dN!'FhrPR9Ii,ʸB8_KqW~m\|0kSO6N`yw@H.G8mY* *Mv 1  ӶPBY\ ^\\;+9e<+FOO[t~)"GvRڒ}k\<`+z3"Q|`<1(sT,͚uasO]Rxb.-9`lra%,"*Z,L*k6pwB]_xl]l,+nD/nuP *vv T U47C2JT1bTS)GQ1p7bt|)ۑ ֝Nċdfs)sD5w  hU*Ot \Wx?9Usf 8*n} !fRzClsٰۿŰ!,;c~o3|Cf9>հ;(+smֲL}lTBp'g旷Ϫ[R 7ݹ| Vm8hvWfo?}1z{}5 NBK)I!*I0u-mף[}wrQ&)~Jb-Vh`JhOr(^-I$o|MLEC(Ljv%,uOqn >Pig yd~hb-+nO 6ƐY.ٌ glz*! n웠9\uТtFskNb5m5XӡͿH_#]TnZ'jPXӵK07;c˥֣9x5!P왃H^!)22Ʈ(.92YKqDA K}Eng-J&ꏡZi<=#Q&aV2  X&Xe9$kZ+E$(aO?U qDK'_t8H۝ !;9bYg@ua ˜t '")%D9\vY& L\C)b c\/6+tuh+DžʹvZ 2 EC,K% a5D^PTdC=WW5f7Q =f+ X.Obj{Up367cfK$ %56c ˉ՛p3ތ]ߝj2or"e%v=SZ <iLaI^ 4/2`SaOʍ$"S6`>ҐjirK %2uύp6/P^8d)ՒM]dJ#ӌ<+" 9x>՘6`YYHrɪbTooFMOQa`W]>Z51TFF]GqF-r9Xm)K4~?6zy'Wbu%tRS葲kM `^ ~K⌫!Zz֢#}WnYr]% hLu%9ntOɃ щF-֟u *Ս\4X !qR}ZڍRɲ>[$D'[^n(vhvkAB"Xқ4DY&0G^|w&&'/1ZO37clI"uG Ѭlz1Kwo16] 굛Y -7t SO5t yÑNĶ=jO> m ^ÂÑC?x&G>H781.#dgrJW,S'e5` OU)S.IZHHR}q7N16`G߾.Y,W!7$ċPʱ@i2 XJ'ۀI6g -?v i۠Bu$(QK)/XkGe4B$B fʇ ϓVW)zҒ* aEiGyR_`iSy1,MJa V!s!R ;\FrVՠ4Gxi z4+[l]M9$VÂ7us~x(b< Imxp:w\β0H'(IMe b7GSۍXڮeN;sZP Аt">ǞSZRb#g$q0C%'DOIYiʙ]܃3Е 9rO#s&m'˃J^y$2kbT ɦoҌUM6s iguހlRy]{kpȁM~ bHobR).[0^O@%*o5zs f*[!tˌ>l@e6nf1=$ژQ̃9`GiԄLvj-lC-EZ?  yu·nXM&R|Se?BWJdBs]b>,> ?Bρ1䒓6SeӉw <b2j);Q #` r`:4 d3 L:I}^ySDeZ *,/zEd\ F}|(gx."7],})2O.Cks@k#0/y6M&X4Mh0|q " ّ9 LK;P'#EGD˻:\{MhzQ#K.͟f?=FX)fH!5mC X"a7:.o>dWf?,5(?~_k?7̝|xx[ƻI<y<"`/tO]>ގ=/_&/9 h}r7S0O0H"#pah7wr34NpHቔIE$2,b Jb,' ʹN2AxiF*gal]{ C=/K.Y??]A_/Y1@@:-|LAXV'&ł)+bs/$H=r_M+y1>Uϳŕ:=~Tgc 7N {k۪wW@ufQ*+esibsuF9 ?6<.y=Q(Iz 2a0I:` Zr~*g,(jШIdL${)όq2"Y-S.w,Ϝb R< W.A(tzrR&SC]=(Q-oM+Y 2ɪ?tH*"U,rv6y%:8Gj:<7˛N2ee'-[# 8^ÇGh]A6Z>\Pݕsx*kcR&Dgq(YYѥZ08} kD8 Z"RW5#Đܯ*eRI0iE^whqzԃmu$&\vӈcz$m׷?Vw:=7e;gx IN,T+>ɼcEn vM1<f* 9^[0_[5|D@H>݇pt'Ѽ9`N4  T6g6/ 2swӎ5vL7mR\q6\onrI u7'䀬Q%;O4ocvYq*=N.{kik<7{kq2Ь`=RwN5-x$pvXxq5tGsmxŚ_V(Э4oe GOў5Ovps.i%!TdV_hS?ǘG>,:%_huO9_y$\1ǩ:#\-ÁPG&OhT4H ;YY@Yٲ`0_SvedE3h$Y>Ř%!1 kJpd*vaҔÝ$,>y#d6#_袓`cF-ٓߏbW( Wxr@}Q^&"T'T֢~T5xT{q 괚Jy^SuRSFcTe;֜W:lޣ`gjG HW 3QdXeR"V$qܤJX"cA f\j-s kN8V p0w>r}ԄWJ)aˊbv5X\/'a;tAR.(ggfꊸn(~ G4A΅SwGnc-;Jvh{"~/k5Wi|Ib,YKBzOផЎ|"%SLwxݨf -:&tcTGnфnmH7.2X0˥/G0#V^ZdgLE'/!?A* "L L\LTxv>uѠ5EE:fӬfD%Je,o7K%vY"Dgܬ\GOll#U'6/W9 (ALS>{>ɚ,A9A#BG)qhpט'ИMx$5fBg"(u)|)-^XɆ:U~,O ' D+&!&N'+?d2#~(u?Ifr9.Wt1Ԅk8E @22' 3xkYbxB+)xS}7EAJT+aCD0Ri* hru>ŴKYe-JHt5>N$' ABXIn6e8>.L< %Kp6!iIjuh`G1lFN%k˖ QXmNh#xx9?k"L#~~a9Wn^tp%Ѧs/EVsAŖ/olڐ~R$;X8FS3g'ڙ9H0kDe  \D:f`9eka˵ŝ.W,1.|^Î9v s"+%5$|%9Sj!h0I;58C0[C.?_H~Y$5 Dw#.qtS0K1wj|1kO F_jRiD O%s)kEiBShҔQL{%Ù1ji#JPBCYt%AY67VPH"[< Sz8dBQ C >&8|S%)8 &JrSNYGiH$OaD 0AŭAǰ?ӕÃY&5G ̷*́A⭁`Lfi[u\@ehG"&aMV*%70n!`A {?,ۀJq0+  w|Ȁ5,eVpjRlK J8ȩi@V'asj\81qYnF~܄(I&eYQ-w!ga}_iө!>QX <5Z:yԳ [>}6}7c>#Bq֒3'qRuTEk,*F8*XfEq+*ye港ٻ6r%WGz;gc|N D ;tܮ=~MUщʲ(BIP+ 6au͞MLeK6BŢlIi\gZbZ%ÿd(qg"5jmxشxN2wCsE8K{dB6ޕflcBoHΒ2R^:2BX{ռ;|$4qת4@">T8\R-j3qbTCp@"i7hrhI$Rj{CN1RTpôFߠ^j՝$ߡE7N+ m<vU[#;v)(c*^LmVN/#؆8e/F˜٭vm>*tVᝊ™}Kw4U4¹Zyy IZi"X۫_^ix0kS MCR|ᑪi҃HɪC %ƺ_7z@׊S1װT,_6/TPؚ|JkQ* v*[\q<&#^lDBYl!`u1PQJ13m"j1uHN^$o}2Ae,vJKFr [qx%ѢxS-WS )ogKdh cA8d\)д5.kϖu>YyÛns G2RsRfAQdΪcv[\ #W* nS h R\F6%jKB xPoZ/-^𥒙*VY4ׂ,7~:1JdW>QH9vҍ:;Nryԥ?Nɼ;~ `hylŏtf6^U"gǭ+$,[盋N *$p23'O TH9Op"$g 1r=ܮ-K )%l-+cAxbg "!f(GiuL)GjlMeB8]zf J`+at BTcFL}rNAg=XZ~~?OzSRp'ܸ%69etă=3rI8ȫsߧQQ-S/2xQ!rzu%^^gE Lҕg~X;)hBTû&pvR?ĕrnQ8^%4mYSp_pXlb狵ݡE):u&o9FnIBSEb&ڇl}j51+c֬NSrb1&X߻jTJv;8#½؇?%AJޝ֨2DɫS"ϞOkkntkZn]5,d}gtN\ūnݿ'm{#~rhr?a2o^̤oMeۿm?(z<2os*s-pIȾ}-H- \ܷG|pg@O]<,˵tC[R 2?F2n?v0?k'?Ï}vф[^RU6*Qq C2fv=MYJUk䷄Ԥ)"K{= %9-Pճ?xZzUi>x7A9nV4tR9 !95I6q#uJ6_Xԃ +݋BD3HdU6˯"MM4jI-E+[MKɕ9i(OtKg@4S ϙ"W@~7|a v"X)E!Jd!RE8g}M\ec)KXb x|jg{ٔ񒇗nn2SsI[loAs_hOrN="KpM2g™x7IIw- [ ?V n7n2kN3UInsnN~s2D߲[Z꬜{&{^m&>IݹWvۯ}G %V>O=ĖsC{^_4[~ ~7!bZovc:KoYo}2?-/k]`>|f/L"` ?CisuxYMMI1 Q߆Zw(|D Z;,~zEz>($>DU~:Q6I4St>o9l !!-2è GQb5n.KGvQ.6)ęz1n jv|M3,ni-@::0T~5?=v'B#wZ ĊRB>@j{ekCADHmdk_ T<w.H!;TqDɟ8 5[u'i" I8kx\Hv>!2]ځ93Dq^A.%;wdS)iEyfv 8$r";82%fh4IJ{/#ԧ1=ONgNp:FΙ]֕;t-"}/e !']Fz@՚hvIt& /.]E[xG@n[)ӂ{xQ !tiMp,&fgx+ĚR!'v'u"pS KBc'kTXEf@-L8TOΪW.ާS")eѣ[m2[.V%"Z&NZS@@̑F{Ό7hMxӂ,FFSLMo \(5URp"z1cAu] bAb@u1k&qBpHwe#1gZA 3+i/eڃ&IG jE}/_ b:g;:fI9cz wR`?Mw@UZNCy b|~^5UO"Jq"O_ÉюHke<>R C>Tu8 ElCD(N}K$-k_>Re/ZǍG8wZ*~1F刈4ٛoc"-@\PG$c|RfkQNUBTzs&S6}HcܔévX7YmokIMey17+ck&lUޅZBhKM <a|[}uv5I?q Qaf0eBIfY*R1J40na2hXaK%Et_0Zo~_%YwRPIbz=^ ᴠOc; WiLڱ`-ZĴ^>9Gl|Z>{b0W Ҟ)8$Kk6n& $ooh#1} 0Ҥgq8U* [wjh 7$uo_Ο2=(AEeD_p:F!1VҜ5XӯGV=@;EJǟ0} QU/`uv,]X\< DZG/#gy \_FsZ`01e5q(0vd90mC%U;qËb"¾9h!:zb6{{gR]Y6O{J=#P-XEZ FM p"6`w4U.+wu1kP*Jq=h8 vƥL'Py9EN91|ʷIQadT"uORp0T 2!THm)Z*yt_v1,h/Y6ပh_|6ia&6L.T"Z&:KR 2LkQ YG6)RԎ"<7y 45p8j &bF@aLӬ0ր pl~d:[ HV?^M4لȔ#ϟ ṛLA8*sdN(ydrN Kq.Ĉ!# @| n<2c5p%7,yr|WL+äa>ZVB]T dR|SGn[0?yr|Uk]LjVoFx_UW#9bk'oL?'ksdbOi'ʽy%;Uv_֯170q2 Aѧ Acbia/Boٚ@xd> .?~ȺWQ2sGѯ._OFx7~ä~nVfnP5SL6cb!dBV)ėY'=}| ~= V}wH2` %ꎉ'Y͝SD(?ʶ_iۼ&I}њmmFhg_Tdw 8tzUU< Nhyzɻ/^Q+I2`q0{%7*7޸-lON;*f'9Ѽ5ȭ)t~:Zg!-Lp7YŬ㏷~s'aI;s'7nElWk[ A9=1@< rdGi뽒=@=6Oѷ+LX#"8K7=1FdQ!'wiɂ^^>m׹)}J@F_Qg8B`=9QS*#ńTnfj$_wf~Dv8f˿b"sH Q3tIZMIg[aEDvO s:$iZ!̓۩ AAe v@;fJ1j1=(_핰]* q.kl/P|6 ŃC>B>\_pr28ɷ rlB{4Upy~mJ˱Im"1j\Z:MnVIt %dVa2ѧ;6AU@C1mwݚ{yaZ|m$J45BsL}iR$`wI=OA:[mr-.]ur7QAD-.O5Ä2<0MޝQ/x:_kqB!$?Ul|]ӝV! *iRwUot} y ̼Wzʇ5XkWS51tӚXS_MkޜtƜnh^I_NNnrI!f*w\ DB°@)q%3 `oJ9y;I#7ya-8 )ݎ9K5ES%"pW,T1B K) JGy :md L~Z{߬U2P4LS??9,óUe?ǫYww;CT (wC$'C\bpw'S$iԼC?ѡ3& vFgDa}3 ƋX Vzс86(cs,PLeT\! ۧ3X[يnQ׭Gl̯¦xs\!$ߟTP!yZannO"bZXXt ~~gL>35~~>3_SYX<)?>=9'QTq)0R>wø㝠*a>QYU)f#ޛKTDjoX(Pd|0 ?u9ȬtYBbHo<&d527gFK1wz3-IA〖T 7S \<&ZRRWhW\|! fRasc\+ ܩbG7\{ghn"W T\m i!#s| #Qd4P\./+A-)mӫDv%|[ϑ="|%%) f*c֙Pt΂J̃{kkɔ]NbE-sèY`E:~sgH&ކ)vj =@oYCݲ<pWya/'[~/|ü&plv~7uټ8ҿ&c6aR}wQ H>R()z^~e1?) (]^+"(0 f׊c߽hŕ˙ZJO)JaV_J:U! y"$S .Dlk7R1>hO%L%4V5!!/\DdRwٶv[p{nĈNju[6Tδ[@c[E4Et,I-h ye0a(y %Y&4Ř e, KKZ]yzC5;5H,z߉6ʅz9Ѕ[IOIF&!hnraAq0 82J=eA+,3{gNo\}BRI$ m;V7s-V215V OX&`8Ӓ ʩA$ R9WϢ&?ERNF [*AxBE˕Ϙ0_~=F sl=Ga2Vhs BxkC\d.ˬQ#eʥ\^B.RQb㮦Q'qӐYD~(2jBB^&ɔԟ%a1>I{ϗ8lbyWy_3=nf_y X\w|b}mcb+6}]0ʈqHOwhƱ(=ю q{ b!Q{.[vV+#kb.B{{K?4?؇Hwfu?xHڽ h6zU۬Rb_QrBbֿz5CԁF5 !/\DdIlk7IAQGϤe:[@c[E4I7OO׊[ ;sz!:# 2Y18y8HK3既D1I<6$\ۗ@D3-h>nMH enutرS{ʲ±0|лS[:Y:;>LD`%S*ǹ\y>"zM'wn8Z%D9isyC5 w > M ^bܴYJ@wTdckGFk||a3?|㹁2 f+;2P*Tͺ{ =0YXeɞ;X lC ԂG_Mno5'O$=bWZq(0f_Yh$r"sߗl_-5bö K&~ŪbUHC8J)ͤB i_o)@ӤY3r-Tf J Z>z~Udÿ'G`bqt}Kտ^0#c:…eu"5$BeSvV/m'7Pp8֌ 'T2@ǦjX:qS'{Yflj6r.ՙ͏ cZHW.o5`Wx+?YZSM^2&~i{M.RBBI~#*d0ii99`Js~SҞD묒~*uH8`1Sj.0% H.밐h_~Hn 'I󫛥m/ F9+w~u>~ bEª݊Uw+xcq!e)IY[)\;7dl8[$y%JRCY6PpـynKZ'Z)JA9 !i"- -5d$e@j.) }Igճa.&o?x9q];"g~,tI\@^>;&T7߃މ2F8-']SGoRS==-\Y wD.)5Ԉ{ p)iɤlA#6}b%EҬk+=_ ]}R4lKJ$Y4,95]k>%gS3:Jf稳t@b-0%2(BQ(?Ko6t5'osӝO]W1Rܿ)y̏jhNގN~KUQzɚEfn~ԌPz.Gyu=~q2ƘY!U]YB8znql^}塀˺6fvnV{uԎr_mE]8_eۚQ>.Bp=(Zq{{ncC@(cLES1w;iO2zR&FOZ>mL49џd|'[3bcJD93F m zIkВ 3NɧP )KXeNzARK+p"JqW <0CA1#[ŕ }XkEb_- f79f61Z޾% /ܞMWǵpۘӻUQ-RH'+,vB\0qOm $͐鵧>f/r7 I\[5z\eNĥ!8Mw"`)`ɺRWJΕ"+JHNXJw Kҧ`靈X K\ط(= CAIVHDUZ6~ EH>9|ؔMcH= 6u CX$5zZBS=!hѬ,N2 b85&"Sr5Md"*=  dިKH|dPuu=J)M4,*atXXUhlI 85CU_Ϳԑ4 LET߻JH\ғLc!G8Ƹ33ʂ:2N0ͨ FT&e$k8lruA]eܲgb--#k+IZ8j\mu_~Ж=)U5z^t3/^= j!VమV!F$ k+t>e ㎠:ڝy6M; z=8/\ g:ȏ뙗!dI mhdP<v9!bq3ujAjCyɄʮyBvS]e$U-i\y_Bo#!w]aUZ0qgo*_z?>twݛߣD:}q>a ڻLʎͯR^T3+kwVs@ĠvWdkg(pQ]|g%$ID!@(/j;.AXAtU|uT1uAy':'#G 2[0sv<Ox!cLB/XH -8Ni J\bB.1fFq;dt fkAxԺ? |tU"~0f)h/OGArBN"0ӑ"I !:轎򎋦FwJ)]:윳,Gs>Vacrq&iw|ȁ儍5x!dm5X0Źa98sAb88qu@|y@F IQm`$\jx>-\`$N kmWP\x|w'L=}?uT+~rmo poުH$ pNӊ7s%Dqi$@0H5PQ*aJ"%Qhi诽'9oY B&ww7b({Z;]'L*eZ5dd/fSm/rf~Q?iPAwZWw/U-[G-CJ#&GATQ%R`15P5|EM/ԁ^OuY7 p7ܘFߖ՘幭 i2ҁRtlN+4.pw9a@B #g 0/D66 B!e"K-n3j6E>O[؟΋\&Xb..f/+[ҁ`i6s1GL5G%+P-56R8T'QA=v:ȼl{#WR"::(pO>Wz~&-h. 㩣xKc3:P8k=k^uޖkv&I vGJ̶ 3u}BJ9yD4iӺh֏F DQ%(j,D tiKGnC裑%L|M:p4c ݌^Ʃ%io?x#x W"'QG/?"V2`2@0,33i ЊYM-uB*E5RUFGy|se Hb)nQ۬x;])H_)InԜrs1 ZJSCc7,W2$6"GmY+cbؘ6H"HRaskGO|yaUZڮ{h+ DXü`^ pD@Gy.dF @W6[UԒL5*׏J sj(KZX-!̨gnHԔF;IWL7}G@FYpʂ`hқP-.uضkv+VfbnDegFi!tAA9gxlBQ@}pvP^hD^0{78!K*8]_Du `N!'O:2©0] 1>. !pNAIS^/eOĘ"$%9!E_RAZѮHpWUU\! sвNq4$H >F@?=V4'2(s.fif9 R\siG 3 ޯBpB* "J>]?S-Y; |!L9HKo@Lu 'nC)y\p<\c ]AA@ހds35xsaBccceA݁jm|C;D*>&H"̺o+9 ]O!iNTIdT.`j)J8d~ivuVmlXTR{nQ 5$ %,dS:0i]&^zU"簶Jf/#D]A9E{biSP2ſFd}&sSd"2$E^C!sY(8 Gk\S;,zhP#Wimt .iM1 fs3"q`Ji&,EX<]%¨F V.)wFC XUa 1jHPDX3+朩ܒx"G] sd8N X 69S;KQ yPÇƻ4t 8"la1Ԁ˯uB^1uFXjmԈ -PU Tx ^Spιe4ArU;az9B.$jh? EfcDmww > x'}ܷ;]@2Z^bhO׉>Y9E sgTrw'pH0aZwvFTIuwFR< 4{n= )TržM1`e$]0ؗt[S1wQ E%LS%xKlwsI])%uC-֠nI~}dIVp1]VߏnntѭlE&U|m5=8ǁ:;)1La#X`*G(6 A©BUCQNxW# 3FI!6O[;SM zR6p`KoGr98A68?*L#8Kx`nY3N*s ,X zoQ{d``}%E'sTc:.߾헩yW {U`R1;?n_^Rg?[ $;OѦl)(=M &M$w p`M]`ҰR_|r)7Sؿg!doٝdx? 0݃H3gܨL2A^\uڰ f*$3ICڃ|0?L٥rXO rʸK۷0a _Iykͦ1?QoA֤TYB߽5L?y3y;<ݰ)~'A?y;OɻY`͑M{~kc{׃=L7?g3'k # lgWv\8DBm6;n2%h[8X{Eu?)OvLreoy:a:ZnZ֒d  3'WiC4vb3s=+ҍ:GT][{΅YAr$kPa]J݅U!PDs8`7Zw^8GikVߚt}* > GMXjQUDH%"bV[eWs֧E'">oDW7?ŃCW1=9\?p/>Ib狱7׉>KwjUy f~⧅S߂z˘͍A DL>}O~7jrړ >I̫&+%eUUUgG|V躁fc=Px[4Ƒ38e *K.l{`GYNȞJ5s'˄.c8P1c33DǭU[),05xO9\57bwfK+ηR9ك BU\s SKXE(c3x0lKŶGSȃ5[ATN1ZP6Zy>mPRD!O5Q"!TD,~> ;PC{|'^ymx8TU̓>5U4ߌç(NgǴw҆_O 9/xAz䘄KOTbAo_vByXe' ן4м=-NFVVb({O]%ڳ ,˕6' -Bt_ R'yQi"/*yQˡH,-ɉIkLփncP Omm󏻟jq!$ܢJWNku.kQT/Ӷ}x\4s)&jHqYvà >in7(n}RkthgF!aET"/M^"|:TIw4L b,Z&1W"* ˑU[ݕ@~EUXk@N5HZ;#B49H1۷P/gkU&wWQyR?0g $e~5@1[u$(߮r7 ,t:݂qBI!4u_Kq \PǓ/ePvf%bRk3tznHȖ-I䵇!M*z})yw{yE21jϞɏcJlnaW%>'s*7w248o'Z`"`[^pGzYd)twUxV_95CX YQDrX?mS!Cu("@0]}w{W/&0+C+(nq)B[/W0aīOݚO]M. x\Gs.4", Ύ"H9 ohu8]4F5!ڐ1#06i'R!{?&v{)5Vٳ?Q% /WWoZ\n1\3@} q$ZùVGfv٥ru), Ŕ*iSZ8kSZ MTs1AP%!\v}TQMnk^G8R{kA j7AWDI VXTT~` \\#J;a=J9b(΁ɉ2Śkv[$jA^}>vE!$>vK>6M3Qi _dUM+fC"R?ҾRnvu-Fr;^g!)pg!H;FY$gz8_!E(b' 3 $ ڄeJ!) }O$d7ot!V$6ԹT:e (Q6bLlX9tL_"i+w ; X`2IJHN V:hK~|B;+9ʴZ&}ųJj t@D-5Š("ӅgSyTdT*eǞnL:6AɱQ HQ!g Z6KV"7|yP(QK0%IJ5ICіI\ˊm,OeU kIgKd5T+ū Xi4Imc0L ^`rOs<-^-WV[&??;{ZWM`~}" eiubuُl-@6ąͯcOmn3uW\7~lp8)^ SO" Tw U*$ q0 gylvCA rFँ/xe²5PX9Z$%M|rrfPIwbdgty;9.JvrH5$ A$؈m,˄9m,(p$032BH}@,xQc(uQPcZ,uRJ+Dr&l#X&/U:lj"Zco`HS%^ c3lJ=UI5Ik[jѬ *h\* wFIc$$%[%wš^1ɮ ֒S RPFl #٭ҩ`1pv$YPpAz9 < ,%Li\I B6U0ā< NZ켎,OId=k*0^G! *jlcm;ohv*dSڪSŵg'  MyGQŏHkFqD^,a ~!v ghIU`[N"Aֹ;[}{r@)F1!:A=@ мJӌBt5RyJ++ N UH]Y螰-x w\@3&ܿeyקKobއ绗W+B5{*t7gȻ0 q1g*3hK8|fg[њK=wE0ZgY rB.,!/Z&3Rg^6SQ-B\! !"0M@Xr)3 aXSkřsr8lsELm=4 $"t{!7J#DV'΂\8ΫC&ָu/Y{b6qkkT 60@qc%8r13H1hBZtu Cx,pF`UkmZa&:;Q_s<=NwoB@NHuBx̐GTe0b %Y@Ajt/r= ECF@EQ ڃn"!ތXmmmH,Լ|p“|~#D$,Ж "PFdVdexv:@ 27ds}e׾+ a @Gf6>(.lWП; xc]SYSaLAnjL1!&z!P:.߾!q|NgƅEr]Jan) DSkʳ%8nAa_I͑2Z \]"]py;fzAnP)LwD( &wA3%Jnr (+d +}iHb$~d]j!ܭ-rr ȭuEM wB~Y/¡7Iwkyf\YWTMIUL牒'#RId<'Hq쓓/ZXICNz}}s@9+X6:w!R)[OfdR! i Q;P}= GL}a{0Vy>sˠGuRN$ a$gbB}vs]xyw?_?nbn:c1d, >]LR浹{I=(kt {W3{o5̜6+Ė&d|^|BwbqC;J!}>CO]Ԙ:>NP[{ +kTB@5gԭ(/5,}rn6yXdn)E l~v9o-l fi sl0[uCp+Kk^J+V=n*ztp^}=Z5b4QЪy4\<ͺ;3ׄ~qe Dz쵓in^o؉5ZDhqqP7nhSIxބg?O)yp(5o;">N7 w2/(s?o`c>OJ[ `X^f ֵPQ'H5d1md,>OSok#8 a6_wkh^A4^JaopwaˍE`[sě_|}Sowb#[ކǞ^g L/0lg0]G)c3OXQnTM2Pe-%&qY{|R-dvgYgH"Ee˳9.?QEr"[+ij>SJqsHM|f/QH٪~LgF:_$'E}wnLYC1EvY)7D4-۞rEtC#N̿jVE2Z+g%^Er[N(ʡEV}a. qIb!zFH)R3B..sU )]$;m.@V-IOHVtsť):W|lDgA1a92I*|* B!l"IpRLS;Q(N%D&KN4 tvA`8C};Klv=!\]mʦ K.qG['Lj܀Q;^f?z;y="lqq!NhrP[y~u⶛DC 8qϽ给fF׭Ԣln?:wœĪ&0KT>mOugwœBt|ZNs&8<׾ӁN'}σmhM]P%k'V֏unݼlAnnctJ͌2,ޓOw13 q$r TznhMg:xPIZ,6TL/uz`k ]ԓMXP3ZE9ܾCy$+=y, 9?S{|#gHPQGIA$?M,N! '$"d MLP2ƿC{wpG3?wU*h}L7cƍ=,Ø(PohvC+ӎqmxy.m=L`wYA#8mnsug@پl/U[4,80Sv )Q$bЖ jsSN" >Q&ϫsPdaj?mzc.-Kc7OTdGZlĚI& KFS$i,͙?Q(t;QdBi빥QP&""38Ƚu42KaJs]W5r[d`3@E``’H 0sf%qkU+kϹB"&t[0_9φ*rر`լ6HmLm߽۫h~8sa$̿%}?۾j' Zzh7UoI$7DG.uX@l#ȍ&O/;ep# *wmmH29hwkv+M2)bBRr u9H"f[4F7`Hs.蜲0C I1ىP."[H$%;0j¨3A"R@C^B}Xc*A'ZHS z3 VhuT-_,4$0JEN#&11&so a߮OPW]$q4G71D h 0MNp >ZB^{jo![>ʈއv,5 nPnnSSMj5 DX![BHC*%Fw0"_71G2pzO+ ] \ז3U!jm"53{*#"+ooFדs}Q8%;eo}׋W(<?%#H151ev@*yq-=e-AXRZ~=xA[Fnˁz(`h U\P"H2@Z1YfrwZQLGyskv;}s:Lp#> J+ ؓa'ۛ ;Wlb8O qrwo?= .J1סfh?޿{Qyoю $[}Zs&藟d:o/.ίN`,0g;v:{ȅ=`{{u-%D?͟{O&_Lhm>Mk!"`:Ρ>Fm$umŲE$qF.DjxƀKhIS2\K y8#\h ~*5M>N!'ydYF_F77.דɛd/Y_8_}"x>;Y&?МҼY]4bd:Ĺ-˽ۛZ{sѫ+IP~74B 7e1FXJW1HqI+ Hډ̮8/b0$xt\e"  1Z!8 x\@1N큈"6Z=x)n6D9[ʃݻmIKVuvsa}<< xJAFHdL9 #VrbdRS>t.k4$ 5짪ˡ).5E\?GoYtfzQwtӼw&ח&jBL5 Lħo! 99ˏ=O:Bg\nCȵW =NӧuF-n1ҭEVa/ofp>9(uWGU HDd];*rVчpK_[*4-kpf1&>#I=%waY1"7q~zM}]|2 x0]; 2`?/ɳ]<١c@:_FWa픔c8IҊ=mNL䕬 H0*aYDA+"܄|E"7HF q-O7@K3$v9Qu1W #,U$y ED1#LZPPL +{- cH*D"6 IaY!R ɂ!ND8A~hٍ ֘NP c+]<2"FCgs bٌia2%.u[X1b|a5@ ,#m1_l2W ?79ܗlQQ|]SSEA^899gFJdZ1#@Sf?]8onO;Gc 3&;?=PJ_?s ?cC7*0Op8m(u; JH_֫+4QB{uȁ .]w1F~\`GsAu]?3P5_IQkI:ҽIB>Ln+"-k+ANBFa;X {u=[(*ҟOE$ӕ֣k38S|1e]/kIPt5$ &XXiL'&wZ-jsBl8Pઓ"4u&#IrA;`O XԈBYс52 ⼃$MUM!l[C:x!|LZw:sIʇGI~yORظoS^X;XNsV Z}n/?mdw/>`re}H-$Oes+wZ$>VގdNl==\_ڋ˸`UfN$q=;9mUr~k2X lMBi*׍˟!,sk>p/> s@l7Ҷi[K->-gR(†dW\/>=8⺜Q9r3oWkXL&IEQHV6ϒWIE"~5l[[J[` DӭKdžsV5CBGbPW0rPg7p8IlOuik~gp n+ʀJɱQ`V3ccRR.JPH/7g8c6Q87Xcf&3c uƧր\O+)/Ajr~׫~MyWX08` N:7nr vn4 q> 1&߮gdz= @K'W,6|`B˞~~n%|ւ, Rpn4譛gSX<Qb4HЛMz 3]-ߴVm=*yq똏٦ ]L|>)LF9ZIR9E\%+ɨ]ec^#"bw}2G̼\7Nu>z_SDq|.'mcb( AoW8q+<,! }RɝU][݇8w}ד_Woxo:N/bu3/^f*4-)5p`m8-FCߋ?lDzlWKaOKZt\|X ntLj;F}sJ1?,; W-݋=pصؙEovr]MW!`Mxю dP9ch噤ja,sʄ{)UB9pkD9.$Z,J aYWwkz2YErX;`)氘wAڲe]oI^uĕMd4i=PE^E)+'!FZn,e/|hzh|_nr^І>($j67+oњ E<16IJ~ݹg` Y`{k>:sS/ »ɘ'D9F2)6dзN}+ ='q4d<@%]PU2'J$Wɜ*z}~r~]IoHLJ`ʀB1hFrLmXA6aڪ4GOs'~~2kQ~v:DnQm({OyYu|v6:;/[lcm˾,`vqrk&͒4ZJ4INPy0sB Sf(%p k$5ȿSVol4RFdI\ m5 j.ثn'$W z[ݲjB (w{"Y=$tB'Hh'4YL(XlYvʲ|w=cRvf^BRS[_z>4Q=+ȷِ7|!uo{׹+7V[;a[OngN.8M7:`V,7 z҂F"CCVʆf_S_ $g; mY-:Zt$kI]YIQ)lH )7 GT6xg,98ӶVYZ8)s %[u"/gIKljI49KwωI_!8]̬q[l8;z%4jhA34Y|TVeV)4xDQC1#_[0#=q65rTmEI XX5gJ כJ̇c͕ФRc%0S`$։%Ň)A/; kRk tH`^)PЌ(H1E9-pzK*iy*gd >]3U0,ʮ@T!ZQ_lfԀw!U [,3A HAÜ:u,# }\L+{)` |„o=Q d Ϩ͂v)tY%"'!C Za0\jP Aw{"(wf,wAYQCbVtY7XN FB3ܨSS!L=t6:TDk@FR{S2 \r0ݼ)Y+8II5T5W-g ¼bx.%n]/ SN1F+|y浆eca`@BW>)+%"|$h'IV תm= JBסvȠԓ2EtD2$mX5p őmN rp=MA@#Ah0kСe*;)PPU;qoXIƕdzv #x9=+`7Sߛ<Wsz>|/ڛ˞֯y;6v|uXh`},8^QL,.^%jWWGzdϱ&߼gɐG݀Q L@0 L+N(UKbĒ3.Xs0/7i_\k'%YL/}X1-"݊Brb[Qp] !!)EEDϨ3wճRv)2m3)%DHtjT#%) QƵAL&8$bFsj3>{qB6& &)nm00 \pؚw-oe& KܡgPC&Qƴ.<" gp_ |dW fO@L>,$P}tZ"&8qLeU⣙^Q/v#W.Я3>VS4\[J۔qhߥ~me 'ܢ)vWZhcLZC{-pr񌓆|_?r&) 3I.auqJ֣I-QhW=U 5@s4 Β"ش$98wP+dBfvP"x]0jKP .jn28H[1icr G;3~bLp`Dkq}Va!a9ʎGPݘtjQrpQQ`=4ղfĄF"Q̪F:f,R % qL%A,|9!kg gsś]ZGC5Xh~N2LɷW oὟl6pw'@*Է'E.PJbB>"VZhR:Sc b0܀^PD7bG<|O0&Y4/p3I޷k{8n-3~y jQ.3Z@3pvc8"E)/-gv,?X8~]K[T0Kbw;^H"DMu<8cyv|CrRaθO6\r +NiZ9I<)^WH)&#h'V+\?6g&_/^xM:vVl~ICxfUB[G7$]X6B9 Nʴ͈ >/QSoɹ4Z5+$,(/ oVafլ$qV+>G:3_RQbxLJ9Z4ۤ+6x`y0kyֿ?գUzoT|?ߢYNn\n PA,Qf IBg,(7w?_x6? <-RaO\ԟ$'^o4Xӿ璼X!4T]HOx \ThHBp=N0XvRq#TSD@4Τ.vx*n<WX=:&7RJ1 kn# hD!-VHrc5Hc*cIC-U"jH@ϩwfTq6 ٖ BB\#8| Sp'͍7+}MqD {n\91O9.^W&B.]gL"u΋vU8,=heČ;63V 2i "6DlPcdB"{Mc#f*!$xƋgޛV ,L7qksu3'qγ([買 l36žQ ίfnj?|C:t(! Cو:dlř؁ 2N ӊ2&xfI:8aFŁ:JF?u*Vjt/|6oq'/P?zynja -Ī 2/?j%")FL "9#kbhyC vbg>=p{8&-*R܌ᣫH~n/V~cl.f &c Cʥ>L@X[¼xȈ$^di)L"B VE` ᫂@r8*kcDdʂJ )2"`J5D:;x4T@/&6FˎۭetoqU@^,ر^ I&/L 3#D*2F*2Ќ[TC2_Dِ+x;ʼnA!3Hƕa,ȌÚ:.rP\*%BxSP-UJKv9\q:@BC4O[dPj>QZB(UlIu;9D+qZ\/YH16`vj( -"B[vK ?XD<-:9*3Y/#WO;d^Lz j[':NT^W֋f mAbzO4'}8 ]Y= tJLR:n;,d4ԑ G-*OC?{പe[C6|whq6mu7-S"8K1RMQ_̾RG*!py_bax>jwrSY{tuH#߉h8I8#kjljZƋ"'U(_ɭ~[D?70Č W(xHa;?ZLl>w>ɻo<9<~r.ު2mK#q?~Cgגĩ!BJ,v &㛛l0<ۋ?Y'/V y0 sbp)|n6i!ݠ+P }\T%RȏMM eW:"3=gO7A:GfCGFpɣM 5EچBBC_gדoK/bl%ōBcPnMwscjTWXf-e4d`ywB]Opĕ{^csc iCNn+!k sB`X&gm x p!BBQz?tM}DFZ SMr_ ; |W,6\9_*qkLIVOڥoP AXnAh;*q[5J+ϳtk\v[ѬUiQGK[zbtKpNB0γO[R !ݾ6YH1✵Ґ*+PN1?D2[Mi[,v>vs?4==Ƴ|[h&NbY{yE)=}F=%)-ݍ6X5 `Dp4ab(ahsXVQ!GQ!GQ!GQ!GE,aT % LF a#.VhBpKe,v*]x!cҚs){)弐͈ANLZg"Z52*Z"a˰9o0́T.Grj%hԤmhuXSXG1 @oGYգu^Udodl*VIbVqef2&ʍ5R`Wk5!+2SF1J By \ "0 #I+|`%HD6X-W2O(: ٙqUqArW?NHA\k{ .JrSΕCgk5bo5/#ڋ^]|^^VskռV<=Z-i`d\0PM3k^Ӫ=-MˋV i%'B*/: mc8zsk80h " P m _ֶD "^kMADb " ":BpC ْ(-n 0Zn)DP 8xFxi&3n+yf-pjy] 9\vlzͼyF?Wos꞊nᖱpUnUngOʭlCL־j+{Eу )OLCk;Ws_[O6Y~& )Z`=T+r2D#V`Ͻ[}Z9s;6ljw*dj^9;c~ٯZۛjYڏ=Pm9nu bzJ XV';Oɯ7T5OTWQ^~~0`}*c('҈~v(-\wU= 1Gg rvmꃮXɃX}DtU} *`Vkw`t~_j[wt~fҳcX93=IW.dVTmvZ#j4hݎwOkˇnݛl7n}HW.dʜǹݴd[W Nwnnj"(Tn'T!!_ɔh(Sb$ 6+=g8}s]OWsxUhyM߿~u\ߎs>Ѷo؜JH.5w!{lQٍq+>eO\[ic~<O]9!Eɒ'?ǜ@eYyEWWӇk.,W0vWSp͢_D$OYW[ܧ20վ-i譐=ǚh\ BTN9#lHt7kϠSr)s}Tp,M}mw4_uv>\R&#x>%]EZQ@:C"w>׵n^߮( +v{w63q"Zb±rgðbЗb#ℬ%o엷DJva~y;>IFHYuAmU̬9Jc4OAh-FS@wLJ@:+ɤ營[Sf2%{m,e.Hڊد偧%Βy˧㗝XR+`u@?-,q2,Ycc2IZxڹ+UZ ª6$0^%G[jlc6h 4.SA3h3j[6|Ae?+!Ȫ ;ѫ_oE'bXn+R$޻'odEB?X8L!D$gTX[,]o_q kٛ nO&W,gsgԗÞ `O /nĂ8~Ǐ4pgC(DYĨRm(*-(vc/85N5y˜,j>hj]\d/GiQX6G#[_Mh\NV!-k!rs1(;prf%=V+3gtbA{Zl_-BH޷޽ChoqN_]+nt_ U˥%{CvL:U3 DžK&v;d1FV%k4؞߯XŁ4:qqBΗsgicpʼnrĺ^Щ !s8[s gd8}kiN5^83_\Y%jת,/}5lWOⶡra. BKeqFHfFJU+Y?L|~ 0P.龗U;)趂Տ[}( ]SfLZ&wOvd#pKn6oGqîEW=ގ}|gQ~q b K]^BzW_Bd|_X ',/cJɍ6Q3lJ<8dZ'Pe1 !Jv1 [cjqH|(#4FoOK6"ǰ Hs$ou_mTTnNJETD6c/se\9AFDĒE_mMzF`&ix^{4 !?vZU)CלhmS솠4H"RX,-' S26b"gTHsl-E_͒SopsH 8חm(:YK]f I׋7r.`Go8՜`80ֳ}[ 'W[}XǪ2!D4f%'gc@ёv`K%;p(# f3, 2rK@7 I) p- xxV* IAC`d16;C(@lr %><6q DϤ $Zi8܇uj{V /+@ ih=J. %:cɄS( ƀ&2f1]Hm3kOcّ{0Zsl7HALW_;#ЌiDAрdwتJI#D43c;59Rl!`@ ʌDd1  `ãKTZv,3[##`<&}"@?WȬ0%LIf ϘV/ ݺm$Jˌ%7O1d46p(H$1(,z+ v97 ݺ4L<] D>*lA.o{n{'FqT (7N0pxI-':f+|ޣMa$Mp !bXȪ8O1}h_z#})0l.h]%a8dHKJ=CJPRPHQ3gl2!/~:s0ʮUgEhd |tII`LV:ypeFqi 7ʎiq2zr ‚"ZM$9d. KM)!87#fFNҖ2v:eu'lxi* V tkLQ(nH(rGcs_VԒQ Q6t As.nDkyZlt(_q~\ʘJh`-cJpi`L,Kδ)b1<^o~VB0wtP+zeB{̣4L@[T4V/jo!4Rjc FUP N|ϟ:BpnW'K[:<m=UXit>ZL๶\@IQm0~X F 0PKeH>JTAR2Jb26RiKɘ@qLȆHp_ QV(}+K,G ҔJLrL^O{肤_& czi986ϝ>wuUh-t BzDXbIHmmXi8-mwNԐ6&9j$ّ t0⌗j=b#J=rĦ6/t2b9,a ~яQF?FecSF߰ZEky-3+Ş14hc4X$Er,O6tA߆.[r+@aOHĆIgŊ-NoHHI[ 'iNdq":w.UxZt*>9]ŹZAI|vBx:]^/N5t8NieEdфOdA>QjVʂF9G߁ ܰbPze]=Ok, $ #Oŏ L \˵[ۥ~z3.LK''wo_f_Ngʻ?YMKdDE7cԴڃԠ1IidhN0FXmb8FFX% +z2'"8HMGKdwӑntb#C`neh^G1;q>: ۵F\4C"Ĭ['_+v1_MN,.VC߇ z1iog+Y9U&;TnF9=]K4?F)qJzoLly1_\+M 7W?vJ-W? 60~-gǔ\ N {}tڔ.[jզ|[N#rB?qh<̜!lQ7A q+V.rG\5wr:|-ԅLpk2]0hJi4Ri/Whb\&7{ԗ_>6F1epHf4/O]>l@ͧd|-@Ğ\NOn~HMT(gPق"Pk=IT3:̮ggߴ+QTDۡI"9-W%-/^:F1FWLL+¿GCԳ܁l,RCx4 W[m<=y5Fwg&X;FmZ\/ӎr) I#<1C\-2[GF3Ű_2^92&bG"7J1FH%^IR9AT*r)NJ u {jj-O0L=^:) K-aN,l9G4mե9!|||~:t! ybvUrbzp#>8:y䁫'O Qj5 )jt-jJk !~ uL#Оz^0 NSbIa6/hu'`AQcsl B*`1/+T#9OA26#(# i#dud߉ %ΓcxBWv.?م~*֬y)-" ajK `ydѫd("WT {Kq/!z4f7MI4Y.Bۻ{}Om گC/Z1j,ukȜBvdt`l0]D$;Mqg.D{CDͶm'tOm~cLx'Nel"8Bе|z'DRܛKDxv sZ/ $dͧ_GC}?FX_yb1q0Q,滻ج}=QC x}[ڂ)_%hs+g?=24k.l$l/W_Yd+ccNafgK#M3 ZhB!:ZP(R 6zs3 ƈ)Rhc+sW)R0f!.Z uM6Ta 64,H1fy 즯 3N󔰼#&,p8t&a ?V`,GFqiۇW xt!JHx v)z}=S bWtѠR>ČbgNL/B-SQRbѝ6AX䈥N1[K"4O 0D;M1̊}B+ 1kN2sB=hL?, |M%^@AO/gWxG28\Syќ I$iO™n x),ypSI``5?F !?m"y',on=*>CuET[әeڨ߿ !"^;}OT@VqN(3rYo/RaXaէ%f1*dyPo)$mn^/o%ǻB2RvHM;QS Rhdj%=kl{8aDoZ֋duQj|ݘVX[oQ ;1U{ָ1jZ8W:-|m39ִk@7e6SZ.#f d XR̎Kz10bVwb "=kai ;3X-p@MdKX$Amڵf:^|ONCI=siS.x#Q-lA&9G+Q|0DmA^}M_濯>w'^8\z1![ONvE|3<'H%g _q/<9iX|8y[ki keA`>I1"scI!~2|~ ,ؔS^I! 6: RʇsYuHR8#2ؘ:̞:+g8,70eu->qf]~Ys +г<g. e.K8_=|VBȃ eƛ7EM^hJ~$]f-ɯ_E:Ba*6U1Qew42Nzlg{{8W SvЂpY'5̈́́@avewiR:8P6¶K`@ۡW8"*wV}3w /nt !M-bw hV5)w-НQ&$h2::X(82~EՆG?twspG*f5ΓcjL0ڳ6[]Ng{Z5E'Qx}˜}h*yAQA^D+BsEyg{;m_ޒ.!6\*jEaUpnYM;iYX߄-[4' O\frB`٭5̙F5{a?IbclȉS1UVYfL @rGcX<yCpēb)e"UF5i:M,i4d(MXVQKI)\]qnoZ-Jve`nm]15_ yZ N])OͫaP):v r-uv <@KD*Ol̸̸tQ*=:F#`1W3WGo6NItėn>uh$zR$ fxc~YΚ|[010;Z!<:x7"𩺶WgOtHmw!NKOzo eB7MC$},ADs_fCt kl%CBUZ׬kZ!tg{܏!a-`n9|7➁ 9-"r@di)9[L1V飌v;btP{sbwfc{'8Boezx7wpnL'pX秇qtqܷJcmA,neq*4)ISHheRly;mW`!Lީ2H\mIc٦8,s'`Yら#>D&?] T[$L؀K˰~znָ?q"lD,t\-՝1fydy#٦)) g_]ysƒ*,}/%8s*b[ZIUb S$MRe # ztI)ຸܺer%J. /^A)Ҭ+UW\ܱRu1ű9+ T]śj,TZ֘r8Ds[i0˿vPEKW<*ֹjm 8sH1f]3y AU 7KH+!9?PvOj6u1U"%l%q -Od.V%6m Nj}:~Y ;4c\FTH|6:*K;ژw9""/P4cU;㓁X%Bac\Kj];:[>g3|\ӌ+(RT`֗Aq)J1"]¿?7cQ=\6rцƘU٨Jd&5!%^K"%JR0j 2eUN۸bf/_j@rxm! 03Qwcޘ֚ 3o@yP'Yc%AyX3ZRW\_i21bb] ˊ2`'gHªbLuhE<`Z2ujТuf!2`Tt:fOVK Rxr7gX[$%bZY{z=bӰ@3WXL1Ehx60IRdIKE17NۘydV`zcё|,+g=Y"Tlj,Cu-ӫ`74i|Jlt4ΛKZ;Yrا% -\ 2cjSnV  d` U۫;j]k.5E~ꪌl֡ GLW\G*T#0%i ,i26fl^ ɱ TaW:nyYH%xDkA]Ӻۣs[M =k`BOhRk7Nj_d%ZDbEZ\>1峚n%boJ D[4!rEU/XF>I/@]=eFn,S7́Ս(W(ם{ BUDg/Htp_{[g7Ep֗ =MFslѠJGB¤LmJњRq`m36'r"+xJ5\3 1J!41Ȝ\o%:ee{,%X.ahOt" ^LGE1ϕ)Wu bP,qR+(¥R :0\eBפ:S!N41D9*_{E 7)qXszcP9L a09디H lKuo,VE._pЉҀ:(DB)/ֿЕ1jb13;}`#vLQpMw6ab(>gg[,< u] cM}kNn#t@FFن pƧC,SĮ8k1>F?wǍ+(H 1@&uہS/|[2zC,xÈP+KRM0z8gdJPX:6TJ:6tpscKLMۊF)J"/NTnA#)2JBXu j0G0v^qd@csQy3ވƕp ]]d;} i~:5qnjtwtlM6pL\ i\ jjʄ&PJUxU^5q66G_I&N}LHqF)I `~j*0\1 |WÏ۝ass?CS.%Y /{:h:}&E/ '^=ֻvkk=iwvw{ohuoL ;yu~?o~$۾tf%G;'o;6FB5.?߲ލmo;'{mğٕT(4& tuyN -> v]0.lUtNl̚N\ğ&ӟLh wx~9I_dUh0n=o~3|ό?ᏠяgMg0£iveǦ @1s/[15Ej\ded\ąٽ/[6|1!c_3j^F6{ ӌeO-SؓWGY?;YzN3ҍoU쟎/lypT41X~59St0Φuhb?vVoCxZ^ëם8wx>I/8oI4Nݽ>͝"G>*B8K@^;q L`^.=6A{xѽa #&z~dCSWGo~3='~'Gc?1[ne5݉01ٳ~O F6S mϣ2Bgڳ[l]ܷ̓1Ժgw{vf3Cf/~ ZkŒH*nh"6nD]&=4|=!7_dNsz]Ft_tI- Yػl PIy@IR&VgI4BS 2\ d14^,ӕG6qJ@p31GQ;2#l#k6`v72\lƴ8"߸A{fwi'_ReR۟P͒'p-+^D/ k}-jq-^3 \(,,pD 8K!#mR+Ra)8w0p:ðƾ\ԁ d u\ԁ:pQ.VXbB؅ m,X2Y0I£.;],3w( 0dJLprJzP&3 @〵$6&b(Z.F09@ I99T,<'GG 5XB j:{v;\P<w*$ IcYPl4< \qwwtCv\T]0HI+YT^ ywI%,}A;V h X@_bNWsXِz/ADNEaHjEkSxrنJ-^.bSY]@ȕV-d̷c2v>n<fPB ISF|jIS0^sC9Ubn ۍW\ǭY2޺NF}f4 ܩܗŚᢈڸb!ZU(5sP0`+ 3&Rn &)?&1 4e8*rW5T$:wKw3:oz:z5%/&7L_:}hͪng~RbkoOܢx* yكVGprGF? 7f?Νv-S `k,rw+㠹xכN(ђ)%غCXY6šzzzLk>CƓ:""V~(\~ XMa r{(‹ (SɠAO%mRfv .m~2e#s_@:/ cjc֕>>̈x NژKF3e*f6^B[-YVnImI_5ǺȻC[2rJISˎ.oDŽ%E#%Dbnb_91΄}]ZŖnn'#4yh غ]';].CiMdi nk8,9߯]_(]_(- 囹Zľ<C hPAѱfec$y`U轆( I0}^ofc? <]?tCyx16ppΛDLT,:7$)H5Z.їL$N>"H OpEA1}5Pp%nNϝ{l\*}IRՐbb:h@։RXQS:-^MdaʚT,&ZN'~Χ#&k,kdJij^ 3֫I(㏖& [Uإˢx:qɱ }24k%L&%/'8CLL^q(D$J'\!ֶFԊǧi,<:xnWvqzIR{J{"dϬ51S:mQ9Dh.J9TBajI}O_$(uJSm>F>IOBCbĕJuL•8u)CA^IiK bNQ^3CEP-4ifz &^I>@4Uoҋ6w8 cZk8h6^ͅ0'1eR0Qҵ0 ZąXLocʪ@$ZG orB Q䘌e|.=xʨAt).%%BDpUMτP&BI9kGNV-0ۂzNJ0Av't?T%W\+rz^,;Zs݇;yv8LDqj0SnǸu-έ } ӊpplr^WdB[!hWрzA2#2Ok/,*$+D6`^hٮJohZzᥗЗryA{1r.h^qk^,E=8_8h*}q'hE/Jpn%U WjznGh5S M'Pl]>xf^9ZuZߟqOnf]U|g$1٪%BX6V9Ό!V}wŁ^:c%;n]-k` yۙy~W-q2Q(rCɽ;Qk ORhp;C=ujN(i!Lܷ+Qp| ;qOpw-r(,okw뼟~rԓ?;EV2+w%F:4Y?vIK LrԺ]RnHּa-6ql^QvK9 {-hsvq2@`:i}S/u2WML~~Uy˛ rtqSk6˳Y-a+(CAe|J,GlZ{<ʧɕ}ɑRKI@n`A{1m*mmv1n{6nk6m{i r|ې6OƷm6=~{8?%[uC5٠k 1_mm{1fYe_AƸ5!=@?_I _"35_MH[~(25wO&8bLWqCm畲J&H;Եcct/^XB*JۉŁ+P͘;=]νJ #LuA"ŕz/3"RN׵x`MIe"ĝ͞KHC3XFEovC#kvh9 U{N/O(=VZS1*UL>(Ce_S\3ߖ hLs"41W韦i%~]‚]{\|kc2F;!®K!3D݌D^Y>5^WSw"lpoC;U\Yl_Qj/%:={ zlҚ} !E[kdU'W)WΒfǵDi☮-/][kvmѮ-ڵ-dڝ`_V^[qeg[: ^["_[731w?A :uVwXviP{FC5}8kp=3kpoQv!;/,D{qөY7Xrg*x1- {PD-S܉P~r~o45a䯺Uy{R]z}LprZmvg(yqFZ|Xg09/\]!t8%G+?]dzWy/~gʠYMߋicDXk Ac0X7X)PÝ;O,E>@TD>M>$4ymp;eڎ8+[Kat.O,hЕ =I)\BO+D‹Vee8*x7 e %DNWČuh 7aZF8LWWwZFh-SejlӏgdsNc[l \m WbVF~k4`X@XB|Pk^T*̔UPaOǒ'Cͨؒ`I&Uk&a 6`Cq9Y$ =&C Qv0emR,eߊmRI&e(e $ŘbNN<g HљqÐ]o:sDK-;7tJ ¸X1iG|,.Ο;Šy _`u.lCѡVXuJ X7]w}&N?M9/i韦ٞyPVDf|:Y)8Ac m|K5#xNHAH bA  k{DOf_:oÒ^xحXk@c  ۯmIqz`c gϔ ,Ou`mXn#s 1zq)/s>)TKbmW)f.KPRݍP50^a86 yn}%BaYxSY./Ax pyFD5 f~ YhP Bhe;Cm-lGCcq`Pl՛ab XW0ǟάA{jVvϤ%<? b XdU*ޅ3I)Bm62tzW3+;2K"[-jAəƐbH3~?t+ uVvp(qHp^;KO˜je ^.2NSMU2\ۦdmJe7 = LPV]ULߙA}ǩ~vzt쑵rG fu3'޿.}x˓D1jAdwװ@50M7.v~Տ7g.O5l:Gu|). YC}xH2+|ͳ,xӼ"A1xnΛF42d$poihڗ#F[F ON1ƅ˃<g ,N x>$i(9STtUԐ*Lj 9U"A9lHFEx:8pS/ONj8p8po/ΉhvՑ`O58Ν׾vw!$쪩t`U80ɷ&rmtύm/)LEf؁%&lc| qRay|Ho肯nwcRԚ͛/PdE/neWBy@SgnUE쌯5S QE8G*ZZd2-5߯ĖVlZU)%E&A>zQA^e统_qdK,] 䖛]XxZ~im=5ΊQ?-"RHd@:쒓oO%l0T`j#k$d(24®(cْ!HF}߬Q'דK)vT'BʕC'GGF#U2@7ʏ.^F$!ȘW Eqkf!gx1Y'A&k;bFXr:*hBGF"ZڨE ^\̟=4l a|E% )O.g;TgAܽ\!(mDN x&S >HxC1(`{Y8Sr8=9/;Dv&e(ZyQGhv\Zգe0VR-ȁ4x h;xI!52 "D)w Fu%DTSV(:)Ml,O#OĂ?4w,+B|b.S*fd]Юl;v:W) `eo Rd:mB%VJ >{KL>MF Va E.{VfWљĖ9w5$VRRMWCHJF)[btH V(t!!iy+P`GN*0?y6Oh23.f]ɘ #UfgƹAAʇ) q.%A l(RE[ZbGga:@M-h@ FIr{O1y mꧫ^jsYOVtW* iQPÂW XHX rYiJLB=,՞5nktֵ-umA[]PKP"CüS_ٚM*PvՙTb<<_B#*2Y )9"Q0:CH$~ Qt&Y=ɽ{䪒:Rrױtdmb%W33M%zop #'S1ڭ) as"m)LNŒUûi~c>Uʆ}4Sx HEw`|KIMăV|j78VXFb*R)oD٤m"@)6$-r{}1bvӷn]hP{u@kA<]DփfZHĸlTnQQ,J =)#eCP@<fvqFfh8NC,>$zj@ 5])Xk2{tmN‰l^LT>J gkJ+@rFGB ؐ?lvvyʄjzqiX9nj{h'rǶǽˇ.^^}__Y$6\Bۿ?GڭCh[o.R}ǫW\69}>^ztMܶϏbq'O{tOe}gwNyyo^vv]=Yoxg)Oz<?nIZA0&Wa]zJa;[e9\.{/63NgV3$9Kw-gjB ̨X`Hո; >W4!WzB׺r@nfqHJ5{%>8"RX܈GبM4jF1N2Q_0sb4]&cg6x3);q<IB f)AwL$!܋&7CR e\vi7 Ń1g,+7pȻ,aRfJ{'`Y7@ J4Io'E{$YD(?*m[FJKVS;ZK^+ޛѠfJXk9PiD?&.wldLsnTR5&a61HKoGJ 7d .'W`ssD~`XLM|P l{@y~Hbx9w$aN"Nj$&-@Ř0|?ѐmO`M< ;mv4퐄aZn+?}OnL&p)h#qN1|oUM|GCڷ}mBK)G@V67:fyM]{vU'GJ%V櫹tEw~Y'g;&`fv[S:#Lmc."lrm?'0 *-F$hౣKa Iƍol+ WNc.}V"c A3N= 3 =0NP \*X5AބGk0ѨvmOL_('!gh`[ #R²W#y]1i%gv9=Ppz#"v5*R:#\].@9&7k: lI#[m`r]ł&g;ndŅW Bx 4mDW dvvێeFJ;=67kӰb VM3fNHa},yZ |f{Ƈrʿ_V9rC+V(f2C)JgaoT1%02!@w/sy$@k2$zr_սM@(8bT5B[*&ՄIZN R+R*!T~\,λSݳPA΄캈 ;-s4F=ޭ_VnD>Httvt +K]AR9#o%(E!讉1\]lxWf\`HKڶM*?_,^ݳBp;O7s]}R#*Z5!첯vZ!ejۿsCIW}+.WA A;?XXI;㝼U^?96r6?gbC~f2\qu绬vcN[:FěX/:'̴ܺJHeITim+^T*_MIZr&jCݍFi{7) JNZ6~ӥָ#"!1|dxcϴh/5!izF"MZ}m:b4=@ePIfH:c5ƈلƶyIKF@nOwxIm˄P^r4"]Ӧ• ige{Rie𷴇A% HggdJv#r<]P eim ѓ9uƟ' .;ۊZ6+1 -Cfi$ܪJ$R! D @pNъͺK|G{zw4:έr(űlnY"4N[A_;ͱ}U73+8dlϛVCuPyTSC@eƑ{ھ&U6*:?@1xСCPnD8ipxe AY0'_Gz0D>D/h0ҌP`BowLR߾k\(ީVN'mYlSu!=ALLցP]Um3#2??5)E6M@Ē;zqrbӊ&l!'Nj <'˴#Ă +rolFصCB-p&Ng9N ʩQ")E.0nܺ&3u.17H3 NJ-%;ctpEGSU *]AhGeJ<pYD'6HeܹwX@+8z,"syU ZR匙Lh`,+m-Xka\( #{ψgCN`nژS{ &PvM [ڂuX M +^tCH]8؁buVsFQ guh0l =-]mo ڠ P*͎ʜM~Vw=1a&j SX#v5ɌTXUڄ; pǝ| Cܹl+xMdN;&Jj6lIGG8ϐ ! 4HSt-G2 OFi-HJLҰnq[NҞ۳$W8*g˒x)y7M8:nc/o I^ogL5k[*Yae=^xM{3qW8w峋rrywuyIqJCiS7//FanFK7zDY:~`<-Y> uBrHz@/%xz;zcM5~K-ʞ"Ͷ:zs2cB>d:aAaQ%qg峋7oK\nBiiM *JwJLM$+ާqHt4'Aos sqdn:8tC5:(ȃѢW)i^frBP\`2J"Ayf` ͊]&n%+s`{3P9tCkCͮjꅍ})K4؈u{j1WtDeZ0}.57`ҢҏWC@PK2NȊ)eJEj`3pdVXt |7OьBB{D׹MoVe%`JXAiNrOCo" RGq+ 8 &Gi|["ݵ_"{-ěw["> :r Fm1f Yqߔc܄ y$En&Fu<*⍊#9欠`|ubé%xE2ʑܙi5q"fF T/E!2Az˔{4љ1BchžREXT4hE ۣ8*(JJv#Kjl`FcKJrM-a܌ N c _kXB༣&r(;|kY6ARQ]A[a= 8izbi5'( a[NgJ'e9Ѹ!`, @璏oA%X A-#(${ *x'|,b!0^cd^@>j`>Al޼{]}g1iLm3 ^8I%Ѻ]'VSc = `SRa͚\Qb8xlHY@&s+\{>9h/)2V˂2䄁垩PQ[=HLѡȲgXJ&""Z! aV3b<ٰPcM2w35-%}Bt ߐF*lgָ{ׂ(6aFuDD0>k5,pCR6bq jNhbu(ChIl/_l̈aP8x(Yֹ&V)Hܬtޚ2P; pjw7)ϺxMOGJcWBYzSz5KE挤Aq|5]u\F 1NAm |+>3i}p,󽮡Wl|'3lvuy9g5ضQgq,?F%n{wcX6[MxS\ǦJw#U QJ<iI `_W(g78hڷ_@Lxy1ߦ7ƸZɨC&*$+OǿuCGC-!9AReZ14|3zxW|QPH٫Rg{K-}uY/opGX3ٻh1{ qwѯɳF&*5^fq})Pj=o'`@;pd'$h '5J̳-r7X"$*1TC Q_w=]3cH g\9QXΝD1(# -YJreYn)tDaetE>¶?%-ڟZsGDSkUqxM9t\Z=//~'?Qhj>5%RVɹ޿{6 `6feEbTgM(~Qm?`;Go($[xV7L%1bռ.|-1AN t<XCT}<dڡ.-1A-W`A5!{5^/@}@wgU\VHX!_8v)[G>W=';t늃&tT[D|,ډ<{s_qv[WNo4n"d0}'ctCp)SlN۸M!;}wrn4߽.Ѯ̓#L5,s..ߺ$|# I1ͷo X| >s,<ȇ*L>p`$>31f+w71Dg_oa,;{W& a\D稲^)7X,˂z9^Bz)r5>=#ʚTvdmS!3xk0&48vCj hꇴ%\Jsqp59WJ<ļT+sd\3Y\ $P`6LۺwJ+E{RiCO&ktHp%1W"Wߕߴ+6XMXl^̕8gW3 gС3D-!eGm:4spS;puMe&ձ&ub1cO¹_:x:Iq:*aer3$XrR1‰轰P[!rf m9ٜuPP L{Pp#ZBgCunɥAB%`+'q$pNX (%wq[w&}S*{Wˢ,$!oB93>5JDi\^,PYKZz"O ?أ kMO.PC5@'13`fm8~l$WԄZInrqx0r?@cFY ݎ[,y'Aceֶ3`2c a#Qcn&"AB\hMARTD9FL3m t85\SW࠶4J~:{cM5s8H* F'3WwvR+{/E;4go>I7+*'n]q:]D%TSu_(/E𔬔gE+4eNJ8Hah"NM+ǖnMn}p΢y)/'-T_N${gz9thu%1,qH7a.QZڮzHك\;P\Yl.9ځf kĹ!4Z\Qbp`X=~%Ef;#IZQ5P# A`[s 12 0?CQ7`5I5LƢòU7`΂iC9~U\q]0&9gaƄjekYlLV*޲`1 ޲% emm[ּN',kZkkjPKOzHl6͐ZJnjr M+/ 0h8ٗ-"[Iu8zZmnue+%YPs7eo5wlфXVzc6n/:Fۯ6nޖ-',ZIVz&K%F֎,QCJaҩn.DQaݕ Dqr6IIWsjH"ɅWDS^obgOz%ʞRD*U5T41?hN}ㅬ}''N=`;Wg9m[ѝݭ0|nnt]o /./ਝN&*M NksOaH_;t\Ywg3#nĈQV0qazT}~~EŻǔڔu&@VgDz/vB(}<-ZJw#cw~IệXusJ}ԭ4׎"cDL03eyv҇9>qiyo2^]F9v7r\9/IKX`r @yYu3˺۞ CHdqgƦ*VR˦{Y!_&2i9,+8RrYKa"c %s&*Xn2b\K~?_}=׺t0{sS}={irU,?Eňng)y!AފEY ]AUYȑcDT4rץ1%n4Ej$mvA8qmU ^cJ|Yd_P:?+PS_^i+Bwd!tQH(d 9 GT9T_Ww^26no?,J&Z:>XrHtŋ3I2XJ]) ֆ;QqϐP\)oNԴ"Dԁ H)zȌ𦦓u|aJ4"a#3S镺`E.Bjl:u@9f Z1lv9}k]:Bwٌ53+3HcV)含"U RL+AjPTEIRgs`oc{sCRfK%/v;YdZ.`J5ۆQ\^͖.c Y/ӛp(H8ʠe>,j['Ac7\^""=}IHC&P7VVzYಾ5+긏BТ؞jg?ϫ7 6iF'+`X41J c:PVEUdsWGgfò{'1 Uoho<% _Cxx)7?%6 jM]iyƔ(1l['~XyB#c۠V01ࡲbmk*W/׎AJ^LgXKfBۊ1JШǐ&D7vYd5?ZJm%Z?}[YkK=ųW9KWw\߬#}H՘6'=U}l>.Y mo>+âivIqVmbl=?NRHRѓyAYI2C@ZJK]Y`$eŝ@qܩۄa,uքyàȳ8OO eWF)N4@zݐWj0>k܍`hfwC@Etz;qq|ߠ}4r`de`Li.qFS7+GغjCP{Ak45#fl%3tP*5׹ڠI|J'/<[\SB)ϔZe,LrV1QQddeQ`)IԹRfBii8U3^zCCV6qQpý>[P2˺.#/eͱRɪ4LA2 (idȲ"5*rSTZsS}r޹ c:t߿(%5O_l4" pnt* T+LsSVY[2Kc8eZrY>ݡ`jrۧ{О}E2Ha@E$+m[-H]=%]L)ߑ,^2saeuzGNe(H::HԞ7Ų!h}zUTVOREڞ=-,MK$ͭw2DŽHDYѢ .v 7enOE\xB^Be~ԙ3E lx#O `| Un_+jwiQi8*^$^/WIm|kOT-g7ý[ЀhڟL_QgsIQ %?]:s=7.?c:H`$:A܍ {4vZA'β`:oUEe?]:b }@?F6[@)\YDym3@U&SX^ԩ):8n,FFڿ6avҾ+cإ lc?Fy !U.!)nސ[rZ{0\w<."[K$L|{KW_?V$'.Ib%u]ZP$L/7 F$@i_D.Kڢqnjm#/mc *N4EA>蕘BkӚ-Kk"(!ƞDʱPTPb"UMS!gE&;M'O˝Wv:)Ϙ&Aao<\^! 6VÄ0~\R CQP[YcNwi{łp;0ERH1qneBIW1쐚%cG88Hbv:zY>kt豕,iIFU6u%1@R P2֭BDH*>"ڠȴdQJhjHekFOZ0@|,~š2Hk+c>\]Py* zb~\Vjl 9Q+sD2O{P.cyߨW?zȍHUx 9~ErXhI NDLZ4I3^>{k9T%Ƃ((oAI&;BH _-ur,w˴nE6k*zGwȐf@/08 GA#&xYWݸzz\-#Vq Y Ā;MdRVY@IEbt4r=GâpYt9Ӣ߰ͮ)[n9Q@%+€|lI5ԉgL>C+[vNmE±T"!E0#Ej o&]$k&FR ~:OWE|:|:]~VECGGsk0E.2ƄD<_j7-$H.]z?YFu,L!p((z;GX&%eѫ"[VlOk.ri×Coaq cD,\W"˗ꭏ!" ~?IO&dS E !}GFwW #rBG)EȬ48f$J*T#2 C59,rֽ@W\L WcA 8 ws4XM[5AXKeo.= )=s2 #ZG{>w'zrwɃϘ 'tC^-u.+kH`you"@9\OZc+r! R0+䝡>XCM&#s(Y? ӵn HvO #}v)`|O*vW1f$u.o'h&s B4~֮ȯ.gx{p:{eV}yv{X;rmZ~…`/We\_vX˳u~w_kon:}aU<$U~2đwx9_|z]& ;D5v:cE,}u Hp~d~;߻Uf·\ɇ D3?|o76^v A=bj+Yn./A_xؿzR@mUW~O@M=z/>?$9cҔ2RkmE>tD I}Yby6 kaPZ}eM4\ER9Q$҂8KOaƷa-SRfpg48\ܪ[ |*b[YF0ѺV6Ô^jj/Pq~Y @a,,XRx9ƟmSv|2S ?-oO~WYupK{=кv*@$% b]c21q0cG#Mnφy~5^Bun@0(ͥ^r =Jl9iN mSc"XBJk)(wTa fRcägJ:@E3vw"̆RI3*O\J/7j[ZcLG<Jx)_y7MtwqONUPfL)Swzӗ .9Fe_xYճ_Np -(v7nu{x77K#ñB*[~s49Op< s g ](MQ7ܽfcoώ~69;əM38 9"=qfų}C9.0C9/ c(Dx;Zt R$2S|BYw%++u\A׽mNw7oEvUr<"r?|`PdARy- =.xZĺlA_=Yek 8k;nc>PLY5 0EBb\xzA01WoHG;V۳pV*RCl1(ʫn C{QSEK{;9 ͕0 a/3_ ]h@*5=ZJ8{G:z$AQ0?Giį^Z m zlnS,LCg x 2zz/]\(QOVƄLV`d!&Jr酧INd*c`%q"aaQ:1$R௩X^@H3%d+i t2LX:M)ɵFBY7* Us/\ɳ; oy!+A'Â4eqgi冹۟~*`W9F]ͧBYzs?WIqq4&j2k C& C\xx}rf}RF̶g`EեuIn1qoA" B#c)a'qJUJM8'闋` t)}p5lq%IQCЮImiWJa#568.-|L ӕ)  {@,{@`=@홁9SUZ(Gs L @`+LX1ǟQ^)ZkLchTT+djׁeT˦}awWB>0߀J)ףW|iH y!,֑]"] :p}Xf#)`x@zڵ S0@<@ <n x/`z0?lLo`&&yi}WaƶXOE+#IlR fD|`P>^2es\Vosd&>깰O.&޷ P/IjV` ߘbmGV`I(J86J umHpO@;0Z}N}=c씵B\sܒr|tt勗fe1Y;_>ɟ?|q@i=sR^:ͷBzd4ۑa-[CԅJq߷ ͺp:z3t|MT fMB-yg6V˯OnW3~>kmA3nɱ/^cMo6؉҄_PJ/ۏm\, VG1WA gQUX1u2NUO svVX' guYaYa`>XKM7ByS6bp@C gQ*AsNd'ޥ=eCJGh/*(LPSlJ5uF[>O&+o#AT6ԜFeyn+:=H%& W"}swA昩K2mп$w&7Վ lVRXR/}tF:r \тw^:=~2|?PU8*x ໬șBUt,]iYvgrzwb2nmGO3[2F)n#ӯ݄qEZYEŠ3}4X֬x!DroGkG飑!c6W W>d#[=&Fk" ŁkCEʼne%΢XcѺ_S9%n174DlM ?8F>0#UޕV߻f ;1CwuYmx9uB{_S-fjBV>u+]rkҵ)@`A1YׯB\mjKT*)RcAIKek N)dC*Xya۩8.Tf[ɼrpSM-F$I{qLB !E3 K ">URHp+KB$H_Xh٫8*ë8* m,~1Q_oqUK XzdgF<`ޞwQU_f b"">9<:YkGbRBL\bbrE_6R *Z2W'b ~QIn2Lު(.PtdP1S0C<勺mD ff~7n2Dƍ'@=[Fo}z+.wP8gp8EÀwn項xDpI/ 3>0q8h:Q^U0G#t[+x˪<7bU[wo|X^^&OCXls5j$Y~>p]GFmIŭڟ_YZ~E2*(m,!Xw>@t_YVGɫ:.!̧:+ޜo8G/ r˅@*F{3 Gv8Bi#(j ϳZû) XŮŇ˦X^n"XF#5hrlD/#w)[J0[ ӥn#mtX{* =2}e'  n,3E1N#9-VyƂs-L?3;vje~~ p_ / ;) qWq+G"?'h@B] PrLau[ZHS6QQ+!G)n櫫4"U2nZ)_;h:4 Fxp FEOӎfk/rv+^K^d$3w{:賻嵐-}YN:jy\?jH?yN*+?|RkcavκNWPd\fvwADz 8s_ZCfx`bFI{d5/0#U < ~Zy8Ǝhόn&Rcfxf@ݴaJ>(::Eo8~/n^\{Ĕ4bm\+uwÇB>2oJx_Uc>ڔe͖>u6Vw1@Zp7余C'7c,HPyec O5hPRSKL;ath{:z7N'ߞ^?9_qˋd$#WCIKUKן\M?ʍSgy0nqEK~ie>%LR&ŕciV?}p.]Ae/ b^^DKb%#cx>9Eu$o3lШp5mF}}aBfXF`7ABFgc ^N=s|gqhj.wgFF&rBNIgFlDA#Bt3>[jKnVvPݐrx;$XSg2XiҐ3}6h3K3eK&!ƫv>{\ ֶ%c+07|Nq)&ua%QF>ڹq)m9!bgŔ`WC)\9̟˛gv?M eWtyFd_NSP_/g TͶyN;%>I;}a;kWحzJ҇|8b{hxQ }g ݜxH^DW4>W&/>}\I:;ezˣ'Ųf:e/} xyCr/ewL\mle#Vr Q@O 2;;Im,W$\$CN\RGc-\ǬIFn-R 'WoOM̧yK];7_"Dm+,.ŨS<E..s1D,HFYuLɳ49&m\hzV{AI5g,F%r:d:5%N8)siC9eE6z}Ȟ3oAT%⨠!)L@i$^q5S# Gt%,q WvBk Q(R,kT*btj1|$ .y4Z7ȥEE͌&h Va­כȬU2KS$?L#pA0Br m7 /+P"8ԉšw/q'5ն0oh*-25ȌO!KMFq&qU[IX)^"'dz t53.ӓt1&ࡹ7_ 4_["oě-P ~9o }r٧G!0ܼ&c#\:==9J(+=9{r6BNGS9f+Gr٩Zr)q4DsU_= oUX:Qi7(hnЯ%x'7H`ehQGU&.K̼tEBt]<,3m f~;x22X:$yd!Ky!:Bz5R/W !9I"UgZF [v&Z%ϰ%X7y"EDTЃ+9_Tf6= O(AIiNTJ,yc-(XKTI,+e,)3 Ks2/*E(*}\ =g!,?AEqHlt]F1#G'ʬ0p,eQD4s*I,+g71HexɺR@׹RV-URp>` b|}56SdܒR9z~y[>  A\t·i`.f~m+ O>l&s3uДVJ~ F|XFsXknrq7#̚-FsYzK45 2rs慹ѡ)Al>}]b .ĝli5:am3 2{tK[ji#D4dMVyH3uO΅&aԽťlsn(ny]qٹՐ`PYsC+5>&G8]oj-6d]oQ#8SCHX8ɭMfB!+Ү*r̺V'\{(Y-։?,ՍhR8ZhKAX%i3tV n{[5m:yvBN۬MtVI 7wč5Xm)D0AH8t†uQBNn:\d榯ڨcX9ZدIa[c-⛒8iZ}U^X+kԘ ^2s2h 0*aե%G[ww ok$IfTEu[BlIݍ=huzU͵ k5'jbtj6mmYdR0:}o^6\%hy+a}݀^/ a}, $@m \DMC oh-SK <}u!هw qgVvH"ہ#&'<գ?aCDgjǍ_Q^2p\UngrJr3~-I]dّ䙝_lZ$P|4Җk/xlt;bnASCLוM8~m%BFZDYn:o({}Y_7 y}2ELG(?ӿU:xp7S7nF/,Yk6x )CBNGu,>C{CHK\s 픶ġ9=cpɕb,`CCN7÷9sbqFX6V/MLaӸB]rgO*ؿhҩ!)zѻ>|4|_]O2gMĔN\3$0ȎdW*k/$\9XXJIR3t¬v8vi)Od2 jY <)<XssX=@鲉}tL*_hci)`5JK<h4tnW 4kqddՑP4blDHW.DT2YQp) i2i'Zp]w6< ņ>NDJiᘶR'=cv!_#;!^4y` GhE'7x(#E ו?2sCXjgm=]eڜ ~m+c&ڱwn>H7VEyِ<wyh:݋vKY:%Ȟi={ͅWvˢH=| }|t|Qmt)}c [WDy]i.pb_*_ňюț v5>Rӑ?q#3RƃHD!hd/my|HBg'1"_wj]JoE zdM휸3y5)o)LLߺϣm6B{?c<_a^?~r+ Ly++vw9yfeV F%1֠k2) Cx7|LKIt0RHt;(|%kQ=hGe(B«!kb6Ŕ \սmы MfW71fR͠Xtg f~Z74sξbzCϕקTƧqX41wӇ' r`Vr/#q!1.Obtqnǡ8bC-d5T=&bc"gf/ңȏ8?.qEO>Vh``pwgh dF.p_%i5o 65έ, r3I{km;J>jMCWņE)Im%eAɔlnZ=*]M$'xx'el16OI47MrG?$x\Zc8~EO գ\LKRdz6_Gen԰#Ļ5M8D5Aj ~+9 8,(*< (C0THJs%Aa/;sS{AgydywtG pk;c!v6ZDcFwU=3C|}0/*#H,ֹCy3D3/fQ~,9ߪ`o|QrQrH#ገdYH ڔq?RF`GMdciI D-Q1[xjAPv=Fm<]n7WH &UaIN |sOUDk1!p0b1mEGlƌ-%4.VqFZM%ii(Zlql3A@3M(Y"':ɲ°7ϋ]E6ʛY{WV)D80 ³(!2ֱqkSw7ͧ_Zm!n4>Mi4,r~*0gn$!O/> F7vP<40׫?l=$ `J՞UKEEH| -LOꈅνZ;|sA yeY2 śIt3 A!\z׿2=kߩG{8Rc0CKG;ޑmّ\NG"לQB&`Z"h8SM.6Ɩ$27ُ|ܐnax6#kӎn? A0;63yg~y06| .jG2&r3}T!%ANC>T eP @]=RvG/{Ao!V!zS@]6ݵrGbFWCW(ۇ/Ꝥ iE?>?b14b2 LV$pat7+C_.@C-yqQײ6xQF>#.&)}-_pvq_.v#/! JA e *~T1j.7ζ`@!ՐhE) 9!Ql4"HŒe:6e+GXlB,H=IWz6ikĻ!I+IKoD(fFĩ)CKW(Y[+H)ퟴr$aG\ H%ktg>krc0ԗ=mMWx;a;3lmu|:LQ=_\0θ/yW,<穦&4;>suIh+Eqԟ e@j/ymGs;gpK&?Gϩm&89ar$Z3~]J1e?-H@<$W$TX%tYi#Li.Mc9㬵BX:b--ׂrVuI&uTBEVɕdc%;vпuHGĩkq k5)8V5%aԈRCsʭbQ8.b8H0?OǣzO |~=x`D2OH\nw׿[ݖ׿2!J;3R OpJxmø[O vԈi{ ($x2LJI]dI)8cRi΀2)ɈQB AISd^ZBZ*MM _F&RF"8:Ĉ̤]Yk1:#TZmudd[>oHrм5ѯfDCtzIb1 t gL)j=o+T2 mE}a7Yo ~z621y=^jo3fTPcDfˢRK4Fbݒ@IPq7lQ* ~yT.  33G{̊pqѨ nk[mka^ ;cJKtV#$'!$7 ˙HP~gth8(ê eIfa%/S.O~t!ӂK4Ph:$o%4 ^Mo4FkȍBclIkqdNcYʨI2D$2FIMKZ1.mhV02CM\^h*{]_$R1 8 b [{^}T+*\?C&zc9[ը`0zו w=FK1|M8̽}>(}8QxpohGGh4c"4 'UPEVkRɭL"+اZxD 8K+: Mb8Вe:hE`e⮊}ȇVnu^7TCouuXއǰ!'2 i\5$ /W;:AXos ]wZy%ȳYa ]+^tR8rA;BPU U<ؼ ;:AV*h"TijjYj)ILI5 Г25՟ֺUGycdEV275PP5=NW5XIWjϰ;8 ,lg0^v/jeRcT(Qz#77Պ#D(ytе *v~.:=D绪]55NӅhweH~݇}/3S ,^_f`2ݒ7OP:b),j[d`DJ f&tMUE5l6P섋NC$A:7Z_UspAgYZRk S&xEPia=7R 1V3␯NTs-ǏsbIh 61`J2:%}w=yr,d~w =uKAR\߼-}v]cn:;ù n|R w<?^x>?Vn~A n$MBzK2uI"^' y"Z%SYҩ-&PF[-%S;FvL;+iꉖj>$䍋2%$i+3x[72jd ,S Vl&watoRma,N' 8D g= #B>2 ey.!^FnRrh<􏝸B-r}v!>.pDw"kq,V2ّȃjuYݒ@1=tKRt,kzϸ"6"ӈ\WgmI1!]I+QEUr,ݠp9/0b.F+W\6M{%hżldi_ a-\>Ykɀ;W4&KNZhPFhg0X;֎ 3\b|2&?V r%0,VT _`x??uX/%eoQL)%@v8GF xDccCetn9v:,iZ! |19L ^qjE%֘@q$ ֳ! S\lZ֒7a2y\ ^髜ʼn,_(18,}>諵P7z@%nMW*1DmA<_K)jPMdX)x5MvGN̳0k)r fA(^BA99=THv(o}mxSw?wvv:eaKv: Mev:#FFa6c8 :oKAadiOT-JԎѥe`TTEJ?yoKaNBbP":ctnOMUOT!!o\Dkde9!ok7)hZ JDv.툻)#iꉖj>$䍋LB0Z=Mg-g[^TD'޹wn=]Wy ,a%t0;˂kT)5N-hpDɔ> 꺑V])Fb|9h[m3Ek4t_`izT"QP`Y˹irV }Y#nBaX(h,,XXf!LW<;X$C뀐prrїn2م19z>GûJoSv@Q]ñQGkƩ6\36ab H0mc F:+#agk3Рl7SƝM-eVeslU rM0 2'|1OWLn>63}1\]o~x5;nX.Ku>M“I#_y33գ+1xaab+cLK·T5|σh:a9,f4~71]`)'mϷ|J81RecqR\)"?V 0w+,`?,Һ/e?2͍_u%ow(tPŪw 9 ^hYLkzByWn]>/i5 pX=QO i. vC1k6tWϮbr链`2S$lb1&1˝ `M֗z:1BF=t߅yoiy+Ks}fv _$Dv u "z `]g}Lrzu'3I 'TVɦ~n'±-6po>lu[!A}X 9Xe>tG/86"R0)iGhu2G6#ɒavI/C B9d=Bjas(-o,t%cнdQ"%(:P v4"λ@JR+AB۳`*c(t<15Y9qذ"`4ZJ~c[ Pb9R RLF۲BR2teSrrorj>ոX2+lrYMܡBs6幬z>I@^*nJ*o;*`jGw~8ݚ,klnj?aገPR#fk,xNqV!Ȕd;1=,I"q,L `w/vZK mΤb"Zr1͚hu[:KYHt!"D;Y5Ӊơڻ@dC M6F`uD|cS0NuUSX'ɝ}T'.hrDWJX[Z;"-!&P39Q(Fxwgd-=?D>?E/dogAAGr"/  o/>&p&X:wT^4^/R*}r=2V)-$Rt5= jlh"K6p qnsabUTr.rc*:fʟ`)[ϙn]H%m%xXN6Y,fGIt XMp%UY-J&Ig1 j1޹),ў(7{]ȀdϦ6lI`AQ4-zTpļ {,Uՙkg!>EmqW a0cBQ؈٘V 3s9?`HIEI%T&Vk -yN'_7ϱ|Pĕ1Ώ1RǘٙJ|L5'ΙB6朖kz \hV.2TEu?EO<UFv.t߰_T6B~tQ){@o.c Bc?4J&9GguzODkYv;UX^}(A`JI"#YB#N2bIB*{qv‰5+j+$sA2/O7)[~3z跗; _ '|jj_/#sD||Ms>?3^_*Lg_o>.O'νξ_<.x>Eے-hޑovc7{W׍JA/L-q_~JnTva-ҭMŻ- d^Cƌ!vg ՇJW8sO]dkcibg$l-HDuO HLbdLVU'+= &Bvv^PB lܻ#\t z_QpW A^VY($G] e]ץ/{:L4&{_boxKo6{k ҪȤ}qcNrnN3ܼgU ׈&ɹLBˇYśR@QmN"䟿Q=m WCZ*[9P->pPHV=g֙+Sw|uSxr,T!АEn1 ¨@\^}x-FI&d4FW?u'S] NSH<"$wB"UCO;3pI0ϡ|qjV_TN0ޓUTI:\tI]=܏&_N ݲ(_|ʅy:__eY ϗmW)Fv=(b5>͌Zܱqw"Jаz0~! ZϪU I*߸w;'H kKa2֛GwD[9Ş6Vݜ:JݼJSFxZӊ͐f]kWCtdZB1i 'K(;K`Fwhfw!t=&XV{E h;cTJ  TN=QN٨XOUαsuiy|L%+FD g"B1Ɣ1,wm9Q!*Ltk*B6\ft1~ZƅQӆތ>-T!ka^ `a6os4Jy5:ʠcm2L+c2Le]ec6:#QX td,1 qO"-[oM6_6L6^[T{V[29|QS`4$XqDˑ$^#M#"&4N " zd\՝`AVDl1@dCy`;43rFAb70cnİRw Ujux2O3"m,\H"&JFB*C2vd[AHct1", 5њ 7` 4ZĜmDm=l_C/ʁBdJW[m9wU7nR$qYtdQ,A!D &JҞzg5EoU됇6\ɵs`;tɭ14So{ 6$5BSCF Z&0 JǠm˷@(.^# 8 mhrJLset_&&vJ/U^4s >a[LjP"`m.U3|C\^PRwIovu;>ssN" E/~/}ۻ3k'ْ_aCإ(PݼVh0A\{Ou.^q0Zo9*mg%noԯqIpk>K4>YȯtJ҄iQ4"]*ncm1Glwk+k@4u-Gc 3r063L5y6]KK &,.j UUGxwUF\b@a̶3N v3[jX;@ $Lҵ}F21P'Ɏ kcCD** ݝbA|.8sA]+; r4L)oݰE `nIkˑ6Fk3;JHyz-@c$7̼ XqI/TL T vbW7BzƩ@c&{{$ y l³Ns ve% vP=^{d?Ը(;h E9=kj3VD߮u .ĭy'i;[wWO!$}qtOΉEN§,MH J d)t#B!ii=~2TD`ƁD&aVEu<DE&:!&4LE> /NJhZyEYahjj,&/"3|d%PkN;zಳg7nT^Yzi . ]ht7 ֣^"Z`!ו!KW޷r-ns*Vn9Oc[3Ҷ{UpjaJ0١1vu`RM;h/e;?`R94"w4ʎwX}(; Pk,& <|-@"޶&e+@Z?u)8߰.9:W sx Z/>`otVR$Ditsvr'zn("aC!DBc-%`\;>z>r,!@2"h&jO<#`/~ KY1$ɺNj?zTY Ѕ" .ktcI]қ<܏&_N_Mim}Z+{lXΌy)U!nA,B9XqzW9R#5avjY N #%M7a$9r+FTR!I3c>9'~HZ`ZEf1Fx㩕(!J RF7Lӓ0BT*:Rm'meY0B J| !N0MFӔ) fzg8^GLĊ21VRT6SAz)E  @0He Dh{AxQU >NJ26_߹;L'ڪTsOR A dr0A(!sIN0IM@Jr/3 ስKBF-Xi.(FNpeB{W8f@WW^,vg=2yi_-ɮHIRRIRRYB`|G|$~U`4+0 PzTcyWLāhو:b\^p߭Vqn{lt+&eMpm, K?ww(cLZfxJSpVo1ƝUwQ7H*e;ݒgt[;7~4vN5vgw 8J*5 MJYX+b h|ҁM Ad`\Ļ7h -{wEa*8Q絛6"FjMS|hHݱOh'ϳD Ns1@.lt178-nACdMgc}sJɮ4^&'.b]jl{^5O}8r(Wt=Uf1o&ΣL#T˛{80N{ܡ|%Z?v0eȞ[X1yA GV*KkBfNNz %], (Dċ]n2&kk9V_(RtdA,ժRWR ip^ &dpʽ)+#aRtxzOZ$ml%YO\nn)>.x酭HU+J!ˠ)/E<⦝6j%Ib6.BQk/z\ Zt GV+ceEYDģ]=0lıLg m`}?V*vhvAj4'V__3k |& 0*WQ*zbNhW}jz\wSell\~C%&~ݔɇß㍟q7]o_߲u'lCq_>u}0'"ogԚY+(Lqg@.?#Fi03뙙RWaRrֳ# y&cSqcջ ) trƻ3вL׷wﶔݺ7nNyC@#?}ݦ4QLz{W{BpئyUrA%+\t윃bF{`8;d\]+k=u -d_kg}7Z Tġ-Hm  =+ąƀZ(ڪܙ,Kvcjald<ǛPƐVr?z2>3=j[PN|8T\wG~2i/xǖ^kG׆^<"g)##[pq܅ºj%,z*fWhZ\_Emj/(N05lMm27xH?% ~hwtW'd8`#/nˑF'@n{w3 ;Dx'|/Vh9VӁֶx IH++\hOQ8iڞcvoUwngw5gM= rBs֝}8 R{vQ1v84m8 * `]m-Eׁ!/\ N wO:e. 8>p!\hOHlF80k+p$DI I*[<˜@0U XgKĪԍhqD9#!T0U XRW T) )h/ kQ>='ڌlq'w=eg)8VBɷl ;bH< mz=3#MDwvsՃį9F< L7!5G8n1c0@hn}eETĊ1}pi蛚E h&fѨWΈ8^vBj߳2udLNQ$EQf?e[YCpct}0A^kLζtddq.>sOF z§p 0y|筄bL^ %U#=t%:-P桲I_Eu**wue6 ~s3d3T-,wY(|,͐SzsvQ| A‘U%-Ky&Q/IYؖF#7gM"pҼT  qʫbXYix~O'ɗ5eDvGD!5N6SC/W@imaD?@ q&%/WfA٤q$&څ ƓP!se,־ڗX^=&EY*=Fu?̙ej@)QKA}~xmVmYޟ[vўcm%,_٢.EDrе$9xAY NU,s #Emr_Ƅ2&ė1!'u-i@:.Pp\[:֕PQKChp8xWœI\\>$Rd{>S}T_3՗LeLu>\iSb9B2Z@HJš` NRTƔ,}Ns. .xfʤ9fZFoZhѝ*SSTH%rD;ŝp2vr*S$wF7JUiMRrԖj+ k+pztpq({Z2Vm֝*Eťɤ(tzLJ@tE* wĀ[%^YI$^?L)~ˍYK?:_Td Y"KguT!T\( O51| Q”܃Up(%*O9SRfxQk2V-Ou{ZzZƏhWeHSWǀJ*kRTOD.YbsB]&[}YvA82ތAŀ4(bVN C*J8VU,RgMMزNtT!)JPei@qL!0!1d [^ 6 25;ZmJoG$*)`-sc6ju1[q_L,DrppoQfmd?:_}g9T7?Ϸ@" 5^.>z:6c|׷%}pE};˟?c'sp~xp__37x<ڥ~~(:"8rex?Rm5ůF:7iMA2 } R@4A@Jonp^"%Uq`7/z\sӸ eۍ!828pDR[u۱^@i~dDk~^Դ8Ʃ2C !`?sB-˛oƎ 8XoGTrqM93]Yoǖ+_w0-rj 6H&yAЫ̘dlO5IQMc}j9Y(ljHCޙrMT\XZ_ͩ 9#$aAYΩZii A\/9/\(D)uڡA`%9Xa۸É*52qn3' S>ÉFE4 \G;_DT!41 aSה T6_UnRt[J=*^zV.-@woJBkB+q6?v6dm_#m)ds{RrY˫׍/:#ɪ޳?\vW)%6’_rUT-1ܛ'Tf{;9NHnq_js^?X!zNƐj̤d,JWLrR{Syop9~zgnF@+5;bvø=f%ϧp/^]u5[?oqzÓg֎Q^qgeΗg}0B& 6 vat)X ّ1 ),;[A~xZyu k ^bo~! &g6 f k6`r#5|rwT 7{K;BzLͧVdW6u-| kThg{#O`C>/OV& GQ7^ Mw^֓Jg9n@3D 1FJ7l\tx=V!uհucwljDMHQZ,)vGæss M<c,;bǨænQA É`rTIo{F0Qw?ܽ#G˝fY;~j20>H;'kHB2 [E+l >Z;թ<(8DϗnAᠪ:+($ tOᾣ| T^bBAr#p4Qc}4'(8Rv(VR"Σs`Xl>Kw#;9 [(ytWlJ#tN8mpbPU[襄+ _GYDh+"HSDj --$ilƘh.RX$$cR̴8I31"sf1O(LkAˌ)%mɄ1IFy.DX ɔDZ_Vqk%ӄȄ~3sn)x9VamJ̿w^jF}I=:3„AzT,#kzVzgJO݄%) 2O!]# N'#'WId9X mt%׌*䢾,μki9R \e j{>kRnzkB>3}|KʌsT;~FI1!`w76ּ#a)͋v皤")8GIi3ɁCBHĐh'"9: B94XmFJIj0)JRrC)ƒP6-%_l^<鹱oO,櫇ݛRޱY^zc|;%ow~[N˵oްk$Vź727g ~Z9qؠw԰k!րgkѹp,;ݬ/JM&FLdwT2?D!T*(K@rP>x퐱6;Oqu;B|#p8!i.߅p=lWBc]UHǘ !3n )AȂpP9[r,hP_42'-n"qّ@5H#ϖI̎@2ɥZ; #_Ίo5n6o\*f2#L%QcNe4b TD[0qMpU;|!iC&yY<߬Olz#U˧fQфo Ewܽ/׎ :d° 'MG@Ze0yޜ1.^]`Ųy-T ;POum ,?͊uoY?S( (1Q2Z5PQ9c>V{| N=M( Tlpk䥮d@]R&knr:mmHB™ fJBoߍjt\c6z?l\~Lii}'bb),W9?׿Ng1I}= $|ɫwvf2)+=?9$*ܢ,0yZ^L\ A5#٭-- /A&[͖f|l_9RcRx$_žArd]hŹTz[ˬ!8,^9'$2 )8qaRNw )K ^PJ xRKI/g2e'ܘ&xu^]6I6` ٘Cuc*/A犳Jiy?n5[/Y岠Ak7E_#R"˗{N|/Ol8WF2!T r6}#*V Vf1{Rp66ʃvLXH)wv{:*!BB hHWjٰD%P0Ke2{~a˶&9w[kAh0VQ|y+5D&1R"Y$cQg0!95|yCX&!a|l39MS4$TqTМ`@&AFL_ˈgOoF)sŨĬ d|^Tj 1ctBu4n,Ƨ:ןǷ88js#Nl :,}$#j_?:Ffc8(LsŨ)d=~ Q Cl2|HMgml<.1Fj:6:+ޣ:,-| D[j\lL "AZ`gػ$4%,^;0gyƊĄ UAf2z +,\M]ݜ-k`0+(z2T 휬RS h)L&h3 #.0!4W Wdp࿚8΂).9݈7!I{ݘ%L3 ;=;} wȚlUC@:4RP"@WtAk1V%oUMd<+dDD{uJ\zO&=偓 '7nBJ*ϩDXf8e@Jc9|6Ր.#** tmY[1*HIɊ|^bʀ$BK꜖DY_[3(&԰ݍ H#E\oZi-L +LcWoU %5 C"$5 tE9 #1~CۇPs}ظȐK2\oQQ,\x4I )ӽ8B xq^:=:QZV hKEԊ)ep[T& KW1 n*N2}\A2}y!]R:lZܭf ֛yH(ˋa Xr0,?/ v?˽߶V*qIh '%<gYb.60::!Ҟs""?Tcq+ŢHY#%ii eF)_"k UZzPf56y`102*G+y(Y1&+-˞7e(}^O)(Xaˊ0'hi4|cCJ7VRC-j CLBۂdDH7JXvk5،N{p=,Aws18Ɇ,ҒZzѷ!R!r qI3#u*[xhEz&r4Gu*Cmx|'M[/ľLθ3z):- 3|,`e{sǟ#&hE R$t+k?&4tg9Rq߇3u)Z ]Z#Fֳ[J5RSr21y6 F!~)afh33#gh 9Π+ U ޴itLKXrch12Y1LV 8 Y)DUheԹh cRew"(3H'͛/AFadŶ2)Et[ v`sݛ@`o-u]/N:$n#i!\S"5Y@{x7q5{dWʯջ/ӫh#H.wh:Mh]J]ᙗx V3=|;O`:豦<`!e~<aofPX?KMRTk`֤+qDE6wU{XGjf9BK,Q)4\cᮯ' >pXc2LJv1?41咠ƓCb.7 Rc3 ܨU5ohHiR2񯧝SKILg5s1Rss{itGQK޵*x"_EpaEل(V>5hњٜ87+&͘Cw}). ^-3UZ< vD'rF-T\u܀Vq+iF#@"@`cH2bI>Q2c܃,,,,* ruJ=ͩ7S+V}^Lmۛ\2_v_L"8Pg{^pʲO?E(E?>?|/]0(N?yp9 B%? Ɠ{ax{yY?WW;v<`8{~L2 ^o \j e7UB ݜ7a-(6m@'yC8Kg k-rK~=杓V +`Cf ZBpF 2+G@{du'@YȺRX]_lҜ"=]U:` _6ūVɋZtjpCF세7LTːaD%xƞ2=R'lgmK^҈I.r"{({+h,<9qGiZ&g3L$KS9@[Auz6( 6H[8nAƅ?jˊ*#~ ^Ǩ Ѽ8%4P )%ih5-E\I`Gi=7(vz ZVw cTehw]*̮1f*N 9ѹC ed٪pSWєݔ |UAQFQ&c<Ύ_2 FkCi]dt\ߞߪLV7X,:rUs- .Fޫ|H&[$WGq97Y :u40m, "YN1biLr@9j< 6j/yWŵM (@5Gi(HAQn޻%Jv脌7`6B=ぉB ),IPh*SH$]% C ps @' QJgP9D(4/&lFzt6F#(mH4`MXQPN׊Y.͘|@Zz] 6Ex6ͣUe8Y8j4T&2듵1xK9oG=`l)0Xe x(Aն(SH^FU#1 4 oUN3,n QHRY"BƤpI |C:e(ԀhsZ 8 mmOX ErU %(G.PƢF"* )Ż &b"' |+l&Ө O2nFB[hsh%)"fUr܃!ٍaN&.O\u2YKI2MRԁs<*Z WhRDYyrhBc$hAdDJ`,v\mYae69Ԁ0f|0gQŰ0JKqIVjIfP/%)]FjI 4qVO@ћՒq\rSph#((bN,}carPrN_ W:6nT PR!<Ѩ9DpODX-$یAoSr,*9AsP[>AlֹA ?qIH\nzBcPRJ"$qFx6;1e;N"bJ%AcgAcB/ԯ^kRq^j]Cu|TV12ǻ>TVX*bm& G4  "5h~W 4=ȯafkXoil l p. 뒊a;ܾʊ L spyE0x;?D?crǴbhh;m8\|sOz/o?x_z?xe&i!Hǘ4/+p֙0)m5aŬdGT$-{_k%>:0WSNԦF|]Iɡ{#/Yw\aoxKl, lyk]ݬV_߰Fa QpHBjzH5(h@˶µ[6Hф5#̻D&\#{_xCMT?c Z]|Ԛ:T3ݹ 藄sƩ@h&Zj@s6I=#'0H,Bt3^Gh%PC/xvļW0Raa1¼3!Z E5`"(Ĥ"kbQA.Nm{>e\DU[mB@:I4Jioׇl=]ߛ~ fJR4NԆ+-"ZY7 @KE[~ ^gA튍c@ƣm`@TCg3j[MJԮ H6*ޱvl=ƮES;&R׻:ҤW{jIiw04>:l,]K_CWMU$ۨoaM3RzG3 3`n䪞!L熱ĀJ 2VsH֤C,{:ia0&;%یAy X|S3A-V~hiKVj֝-x_wfi JRC.7b*>(RUOay?8A \) I BXŘC"*J"wxG0X (iI.)Myk6BRPjr-b_o}VG"#Uw!9F+aub3 {>]( $|ڍJ:>|%McCM-HgޗRUjX/ ZiX",)_f%:b:rBuڸ"ֵ N7l&ʋWIwHK8q?&ױ;hA 8QLqzjPATI Iu0;`)Dܽz UQjE`cϝWDYOHYg9s^c4u:1fkV\1,IдI4ح6y@Lfųi N7fyix5*UI2(4I*]e_ C0Y[e칵V-5LuU5KV~! S`t;n\:7[T0\HܕEZT}:oa͂}.D:&:5VtZG'R#̗Ì wqjR;ȵ)?{l uuCZ -Ǥp Ț K, JYgmD%rW;2n'^J|x|JzVeIaFff/.I,ћyu؛Jn׋Q7%1Iw/#.K DMȐMԊ +*XPeMB@)j5Ǵ}jg}We(B&S'^ DҤMR2z~^MzK.}ȨΫICDdI : )`V#Y6Zs/eud?})laV=1=SrՋ=\Th*)[{Yr*)Qk_BC+jH U_9n/T9IGMR7x )K"<#慐lD%$;viZ=&0$՝؂ݢL󸋗@zMVSZ7hحl%6H,!C8c5k[5->ض' ߱@$m%K>>DfGDfBc^L)jVyR Oɯmy_3Mk| .$:m=z/Pqخ?hgkJm50E{Op*WFsAM3ZtKk{/?Y2&^oRÀH̡é-=קre8ﺎS p8v =[)ֱ0_Z0cʵVr a\Λ7W䀘Sb"ya\}w99LNL/KrYHc1)o3sFv;bC}YfE 1åbuG itH\M{P6ݭ,hy=*^k$楪Ңlpk Үf2ABb)#i_⹂aR\q?x70J c*"p(oq,0%kX@ַ&Q̘<+Reɡq _Uj$Fr$N;L|l]Uf0ڳ}'z@+[}Gȷ5kԣ'7̪"d;EBQ(4&ؒŞIw?C7xFA`,I}~3Xr5yqɳl j6@D7Wq>]OpSKyuӨ\Xu:ý=uvoz`[y!X>"`r.w{֭E7yQ$Ӣ"<:ݧN#@iyPakALvk (au)ca<B>*L0{9b.ڂ+)< \Jhx | *5T+P&D:>[#`6gugtә.adڧ1$9$T> />2#@e]!K^d +z|ͨlǽ j̚{i0DUMvoAF6+_$T=72$Ho<A"DȤPmo7QG N ¥YhڠqI 3Lͪ524KH}w[Ӭmzb-i8+f%(-в:ÑoxRI-a-hųhz vY1?9f]fq"mFMN4V!5 _b?[7YZPt>4sBI6 ŵ&)^T0^}kL=}O *Z%utf@E&zz?8La A׀< E*h!MT )}m/Gy;`Dhzy;aB63JI ?-(4[e7Tj08ê ]VF5A;1-:ôGOqN>k&O8;K)yT,ʛ'cgug.V_tΣnɳ= y.=9^Uﰽ^Y`wGq2gv?L !da%^~v=sy;kApT#Vł6D!W0@5-Ƹfjm)U yHȾRNQ2/hsD4Fܐ1Q MTk\@JC:(1"ۂ7!oC= yk[CV!s"#oAԷx@E"oƒd/ySտFq7b9h\-*s} ۔5p3Q8(F^Ե/q{I?r`[ *,@]ׯra9>,(l3&C!c sQl eh6Ѹwsm` ..t( ]\8hON;k7SN= y[DN3Mr0)U@]m|Ʒ;ߎ߮V};&#xSO|;l)~(Mq<r?|;(ql|CB߂ʛ8S*CGREm9a;FiyNk2xڞGl竫;Mde!$o iDf'.}pOSUT4:M4<{ʉ2x,۰zY8q^v 5#@Br޹,MbrG{Af) vqW2kuZ$]๻BWV*QR f 9XN"PhU<@@R7z#I<5T=x$’vVF]-%k Md}B tuw֗UyUUό\[MI.&N,qsY)q&`[J\G/!v;Zvbz _k3-ܴр`9@0/8w|ҋ7N'8&#㼒3Rr_.):@lq;? 4F3!J$xS9H׆N{ʰDt`oxwJJs)XoM(I%Ne8x>Ğuf&itMF1H5{yUmWRW';ї5$IuǥӉ>MM9'RVs,ij&row8Fi !G$slݹMi8ZI,vʁ&ew\;I;$yu?yO?ǑIw:y))ラʬl9wq e&qXm2GK[[r4q*GK2(GJ?^t.G_M;;hxeY%ꊅGn8gBTȽMG; XHҺ;ؓ{ٸmjq΍eG9|)$0VgƤ2kYLx g맑IAuΓl6'dL2I1عXIwʇqóI01T,b^0LTv]]o9@ݗVPpAVQ~M( I ͨ6 ?NzUpx"v]FBsHU І88ushv:LjQn0qULȉ1~[F} c[]m h&xj ikHK5u5^7?^~E/|*uW^U"mͳ~/׉_ŪEŋghtFIT Lrou,$ bAl'YX7]xe13I%+:;ҔRʩwp}K*9EK{9#5?`IRSG \)g 1ֱA!cĉXE\a`{=蒖B1Zze D9bZq.0XJy=XGs #u*u~,GAh'"qğ.NB"ӊP|yA&'HyR*]Mn^gfXÀa۔gz <\TEX%^`<|+'lȅ//jpJ> ~񳂛A%2񂇋6JqK0tnSvzY*6/? ~Wf}b\o}oc&o߹lQN.1wϗCfh1LƲ|,囷i޼\zqmF|}cR~~AzF?{fm,̇QX~=C*JbG',We2S>HVj[&0k" Go1{íXRl V ;-XF*u߮/A@u uq]&ww\tI걠\m83E)L>5L9 h cX"c2$e #.J_%~:՗Ts+VpۻJ?+VPx9 1!*92}M{a2ׯǛeO/"'2g7 ?ZWڵx4s j'o T05 ˏ RҜbco'?#]FfӏZ5Z H qXP mٜ8fsϊs%v19Iprb3r_*FN9 ~̰0&(E_.ы@Dͳt Ť_f~}Grn 3C5̢r`貱ŠWY @_0U\W 8$8mu2W#@ .' -] q;jmǹr) rcb&<\G9@[$Ak)SbXr ͆
\sfIvX3;ŕyBNZyFl #zİB)`,0޽#+CsR=j")]OŴ4tSP/w^Γv`?_VV w>-COyB#ך74#u> MtUK6ou> ;O U پcذq,شspyوV`=DH\sj \Tg7F͵Rr8A!gZ%ym2"F*lup/k}ɀN~KbĿ-Jvbz _k3-ܴр`9@L^OzrC .hGѦ| (Cw6)G0tJ!Ɯ?.`~?"/P(R?[3 V_RWLq&rRy\MY>ͮ'Q(7aܸ`,X7AL R.`I J(wO^?.z\[Kb\xCy!֛,*_;{ap}?>Ƈqһ `@zok_)bVYC[+AR0ө+5"0|.Y3-x+,<>꾨Ta^;ꭑ< C_34ɔ"4b$0_v޾-b-"p5EA81\r59gÎR!iEEs=!÷YqʢͪB6 b|ku?O7;D T%eN@wD\48N%Fg@y0<) [09FkBBq~l8؀CR&ɬH;sDr\I`6z<7`zo51"-CN"ʭ$P 8MbC-]]]]Vw)u#HyEWqʳ[^$JOyٺoH|@@a/>ɊJJ,vdxE "2eG !{etlX2!Vu Brw0r WQ|~}+v1H!8L QD0 ~ޟncxT׮@S̐z(&" ~iN~ ]Hr&1?oJU2֫:bF5yoZ2oÇMAd'N|g -S00a+ab VOj]~ -"m0tWvQIpkn呼q'=8_a"=jEUM嫚P)d4U -Tkۃt:c&mq+[M;p;hh,b= P/KV=Z/ ,J"&D;\4޺KMpB }0HH,jp8Lr YìӆYL|&"JiV۫GVTyS LN=EU3"xn+]NR*:1$nmeP`4p&$1zmszyP)3Vri1v+.Bc!_=8@Nj`sx*ӊ<ţ) ^3[0K !WbA8/HŝYuS_Wd|oe_dQF2[ȝolݻ:L>Ř0uTsCackHQZR4iKXKɃDCJRP0J E9ײ秶pd*'`4&!:W DM-B0Tc!=2r 9 ,"O{^G2 6jD>ay8Pd3IA!g=aU>Am$!@^H]$H>WqV\{ą=`0DoQ4@*n'*@x(K&ZTy/w^ZFXi !V@N$.>Jm07grh7 fQhDe<yunvÄ|J~>8+ҽJ.S_݈\K<8IVݦˆYdɻ*# ̾VB9;>$i' &*+j>;tP36M) 2a%2V|tz1Gzu=ʃxwAqJ/j*F.;'ݱӉvϟO&^6>mxؑxlXD;_zK~z f`d/yAI$_~L`|} pqogS;md`_vKIl]}էo^YvPЍc| *wcd/F5>L.LFhEe^VNGXLeh5rT$Wl3s_B$Gw΢CUP&AtU`CD%+عjWP)SY0w1>UDQkRSQBHTQUE<>c2tVz5Jǧ!A m{?=I'M?_Su=qN:mR0SJm6JnBi]M6v6E.3#;O4ۗdpTmƖ{CG:*?ɐ+Eun4)m, %[*f@ctuN EVz}GWy%o-l鱂̅T2¾2Jg[N(+Xkd{TTZO,qTh'x[oyCU;tx8IВ8:7iŬvSB/]$X#oڀ4ִFqkPo{s_!`ʶ 'z=jNQ EHk#SA Rbdi#tRqsv {&͜Ia 0 >y迴,nUԳߥ~ǘ~C c*N8=QSBQH~dҨ@35M2/x+v^hPƋ AObsnҜz"L7t H dBtu n"Ե_>1^\f{DA${YPJTשBh\vZ`M&WVÐS:Gos2j&!>$x!=՞C2ǚ߅@ui[UsR爗%Kr`cL?{ iUiImLmsii-Cxx%򚃧^|s g0,$JbmP7غݡvIL׎v?Ȗ6se)V~[ZHm\-7%a)C'Hl-B 9/;ZXCަqqܿTzX%IzAȑu_`rs}ZFodz'/aJ\r`ݮ7P[r./䟏/:szj_=;fq Y1lwrrD04ʒZϹ^9i/-'\(ߧDlob۟c?)&Gh`2:rRh8o:pebZ\Î4owZ!L2Z+1sU뗡V+-9WV ) Zu Lt6h/)hQjOpsWzwJ#z vάR*2x94bme}>ڢG<‚6"y~nrаBPUz-~okF|# 6__j4w5/gT.IIКچX13eBh# >򡺭=~Qı&灓w/8oלM0Ǯˇ&!גI ~[q6?|Q/l?Ɍ6 1^iLSfv`n8S(Xm#v{,L%Wbu~j5Ju@˧Bnc4l;P+,̒;FC]0LVhՕ8h#tM5"rnOi-_"B, ˗+ad:yqb 1%*Z%7gru`i>sn8%ϧꇛzN*& "}N!-ՔcnRM%7s Qmx+fҖģBFT|G8D-Le-1 wS~{{LjF,!w$pd Pӈ\Nٌ!,q!4-Fl 摠$7[Nu)`e`\L,4\^i3ESy=V)sDP )XSDERJ ݦV{sF8oTB2Y2*FkP*FcA8F֡AqĶz BQӄm@o[Me^.- )B/e@K4JK}GniPt6Y ̮"X:F"x97^ Abz$wVʨ8٨*hBR4sBT2 {A`U=O;c.š6i˽O-뛏8˭_Ae KRl"x܀rSmLUc*Elɜu!Dlu{2S-"[Ey >B,\z7PY$f9Nd`CmkCK&ڃg1*U4гN1t#GEҟ#` `:"}@ӦٗoS{iʢ`9]'u>6&$÷AHijUOMCz\-z3aMvX|bchlm+@[svqHLJȬ `~Ux$`TƶUbj el@%U-52d𱜭pu-i(vV5R")sZkxkqxv$okё O G[n^ӝpRC]}:rS9$;uأy;ROλ6MLuqH6[NJZ{;ƛB͕P@,DdfPWYe0e"kcbw%clt[Gm-n[E"%9S";dQQzuw*fr&u>Bwٷ'UڢPG$vRn fgl׀ٙLKbtּfvƚ-@8xwxճnH}5h`Qsۉ3F̑eA-Z$zP(TG uEO$8ZiXL moopusqR4lN3t8en2vB%ti UFFRT:XDAe{ `f߃n9~3JGR+6yC nE {P|k`jE¼{Z"ɃB8`9n% )d;6'dh)e-_T 1Z Ic#Qx@-RV BΪ2yh5Qt)@d RNlQ% N z} #䢵UKi-X!NQ.һ_{E5Sg b7Xi[!$rPrHi(esKw8cˁӶh)^!'n8Jث_h\Y\ 4teWJrd:/oqE]Newx $iqr 1˳-[[ep0tY`y[ pu%bH,D. 7y; N4Ie襤e3(NSj}dHZ5Kxt{%XAio_]A=}E^' "[o`<"MCZJ':]}vn|uꢋW1,L|bǫg'݌4*z0NZq9|DNzPNB 9Xϓv08ɚ&E|Z0N48hSncڶzm;HCETVN GT|UIQS{F4o7q[0J" {nGDVIL P Vס |)%TZY/8B[Fڪ?tG=ذo?īTφKw=!F!igN':޵q$Bm;T_o} }H_?CJ!9Cr(Rb;8쮩]d7vJhȯv"ZR*SFnD풻:S/o do{QN@h}+(?N$TGˊƫl&CL^VV\ٲ ouhS/RI_1]&jD|,dh&>]^ pu*F7h65$> zҜz(ՉH n.0r3P SZ;>5Z=!ĉq8n7wE dW.L-6>4[svh`sf\1QF@߅. T܀z5]蝪1j#&dQ73k{z'3_KV8)>ƀYq\[ʻӇyroh ZqԻ-VmW^vp^6ƻM`!DL(ӅISmYԘzazK*ʼnh,2dkQ DF 8M^wX%s)fjrM炵ߝ;ٌe 96xz|BD=:0;@ǘcksNUCE?6x&/t@|i 99%ݳ?o?kȋ"03!bsaQVA]&9(0_rcAv&~~JI/_|6.;lqF!\کSX'] `{nKbZ){IM-;gKn.!(yfhRap8o1t𛨬;piRz_(iHI K!yo?N!UgJFh/MMh_gH=y(!+*m] M};T s:PRĚ~{QO E ^ 5zCm8É|| F`\ ҕuQb0~(WMK_>:!&; ~.+Ү=7QeϨ?$ԈBSNN%ZyAZlDD% A(pNlAjX_x!!ih@o&o1iC^* ;A϶l /]E} }]v^!rKDR*ê$~W6iڀV& '<(dL[>svFr8u$"m4޳FFH'&WzAD]HMR/X#| 2@CU:?_)8hMytZx(;A vu%̊9^TF@7ԚF=Ӊ!ZIFicכYvA>fio 9@oמ!ThS 6/ )%AXj߬KahgxÔϗ8 ⳫRL{77>+ $5Gxvz--ߤ7L#W; Z$"5 !,q*l7ѓ9b@:X[K=Q938|F l{r{?ORM\5A,姟~J02/Cڻ8>C,-fo*:%O_@%7㸼-?E_Nwnpr)>}|>XE[FR avo~?#epc(p.#y{ %ջ .@ 5 ToVX k0|{ =m"ʃGd L3YvGCųCGVIYkl]R~$\+8jVrbZ<^*8:*C` _گ崰2+?Gq0'Uʔ5~{/t"x;gTpoh?¿Q$M5O!#@N>R9ȶ}sp9U .okRB̝ۖ|˕"eqauBĀ !F~ 7|a/'zhni2L_WЀrN:e7uW'~ƍdZʙ6cVٱfv1 *mٗQo DSc3IW7M=&= dΟ^Jyg9QhIP*.) (W5q̵Dv*" JHnxDXX{D+TsA1/ϔ?aOpKBkdh^+|%Ӈ_%$t&teC>LYwv6 tzL f#PslT?c0|>^ov''iO=8NRid{A܏GŏqAh2%/}+ }z3O'яC*1xiƊTj&;3R{߲ڸxdZDzuДcE ʑPC}(4z-fI{.*G)ǷC@5K,O1jE2ΠЂIa V-̎%FPs!KeG/NCmoTӈh䈖^*jz#X:Fzq2bx֒ya 1J[PγPڣ=k*IqJphg&n/g{$_uEI&*U%O4(QG\QXs%xTmbI3NJAs|" H(WܭC.N 翎D_g??پ,i)pwOr$gcٰrXi檾̹~YB~Aw]tW0ߗbF 9K{@ E| $~} . Ǜ0dދh-#C_.ys?K9).Cu0~Ԣs,бéaʉD0`Zi:jw_e4G; /޸{;vҊNsV+q6Nvtsuq.o{r:e^TN\N9@\^+ '|@"~T"o6. suFu&do,m}n#BȈc=_ uO 5/z1CSy\ T~F)z!}J@iedX0NТ<|S;қ}͐c%JQtC6#s]ɣi h[6p(qllhg7e 8ms) tU9FfsO@55Xza3AK FK(.j5hRMhfV/ΩƾJ2K '1#\u}'F5ޕy(M& FuE 5e+T\e%BEHd_fE'D.чӠL_7n嘊kh=ws5sQnUL? ٵ z:wv|zzXMXW #?V֧0 m|>1uz6H QSs܆ G)|@bLؔxXo>6vR4U!*KLdISt؉}haNeƣb0"*94/.{ }pV bTndIR{ D*_SϵDdH48"qT 3%@^KBxB#]&3yHt3ഥ9`Cs{QIJ8!-=ӄ܎r';݉}aZSlH4#zo^FM:&y8'䁫==Jt#Ơ1G'R Nk^޴ 4QuVHRPQ2C;1h߉5XҤU넬촘uY8 G7]BҒYy>V ʄV\j!:YV, !'DK񗆽qHmr% Xqq%;ɫrBPj.<C.#4U 4ņPaP]0tUZu_Q{,a}E!"1uYhXu6kph]zWdX0e+\dRz$h"=ɪH \55cY%R "#W[U`N7_I#2 { rSf^\MvQbmr]""Ak>;b ;Ƒh z =Fu:o rMAV(0$M`Yק<*>vE:A&޵E/~? vA|مnz&ƏjR3CXH$`,&EjBh CUzp57" CECDc,q+w.=)s]x! M:5D呉+(QdkU=;QM3cM3|ИKiq7j2gǧFa]1Ơ!u~QWt.卉+0 U*͓.udນ٭&9^9m +!ct']([`x~qAeᨋ &c_'TQP3{-'MTiGu|Z1*ٵqY\-E&7svR5n>|gL|0yzsWi |{ )E6?ƻQ W1huw~̣R|B;?d1ec|>J!GVa &~aZu!oEZ !,q&:l6.t(۳3N]>Cfa7w׫fҕ;T6#?6.[QqK?**?jc}~Q 8tͼFeubǵpY.5A_NDJIuJ 4%  a@(:o FvS6{8)j2]xn|{횼É(6$u T{W;j0Mބ5Ke8L!2aVLgpEL #aF2+6P9~ S׫6&6Z 4:/=6B۫iݡ‘िMcћսʻhkGTxss99X06ƼJ-YqO$\!ygqOظ4D0==3oޗ4F`I' X56s(H wol%8ݚFҧ8M)0O:;v}'4C(@+(C`RO6%*KcL J5 <e(uxFfoש)^4iog3Xp͟Ffq"kqMnۏowYbd1 rLZy!wI>Y8 {5za[2Q_!_vѿhu3Dž(q, )N:&S`_,W7/͡R,˝yGkv42AW˞(Wjf= .6)9 UbiJ[eOÝ[`y~12/[EMVrrV'E3} Ԥ{|T׬sRv#6mņSA?>}]"X@Ԓ.i&e$Jю{9X'qQv][p 70)`9psguͦwidEՋGH~6|r{,SфB,I>2*jB/Y;ҺD%hTAfVf?r22JbfF&qLCg);Rgkf[ϋ£n0Ab(IOH$XIҦ!i6P1ٙ!tXB~LY( 'nQl㼘O / GDlyxqlg+N?mMˆRi3ǓE[NFcx/ǓBl[8k)@D%PfJSmAw[U19;`v22x~ @ S#ۿ-kgX؜(TU<[:RD?=/_ϔ9R1O)N]jbo7dxt2IRֽ{q>΀_UK T{ ~: c2pQpŔ-~qcהEґv sh)c{($&i Y?Ǔʤ\vVgNf%TW+JvĕK\jv\pqb%G.n+.ϻqgIexYm\?#m{NBMoEP\CI+NHLJ5:!8t\aUKT'R<ÝFpe3kB)yBpƙ͌ɰ e3|_i!=b rxKNzN&5v$P1:`Rf}B/&n<xa>ABꮱCG<1+8SZPIGotz}v |']`xh);` $zJ^%ԹڱghؒUy^UԊS'V-cEB'V&R s54'/:6r^\#$FC9wV6X Cn3w/T b3kMl7_ҫ>8%yƙiʑc25To;c>eB#ƍD)3Bҭ>Вr $p|?`TC˱y/U-l]}T[jjq6Y ;iJɽssT>vܫDA 3 qn3'eCӇ|_TP6`y/5ō7{^9x/~,z};E'olƋ \+kxB#)OTFI1ԟLdqUgEoI*1,uf\ ;T*iR R'830r'[p +!w6_Ohh}v9O_xĚk/H8f\Ơ5Fim5XLq%JP sº ;NS/>jd~A_ /=#.T0֡ yFI2MX)E)<519,ׂc 7CT$Ty 0H 1p8jO3@@*"a$*Q5@#%} @fA] f&KLʛR,DSХ&)cxMk9mɛd;;;?{c38⊴ׅs/V,%Df2rxijBI\Ք~=|ޠGRί)3t2F42+S"u`j)ϔeF0ưY~s,}[ʹf`HzAju֒,sqέ䨷)#eZLB(2R# tKUD Q1?1~пy(P֗WC5_S✅^@QT[W ] _}ݿ"ᤜ2GF? ܜ|wѫ+'O')sk拿,ݣ{K~|L@.#0SݣWĘPS:?w5l>:"xDZߚs/Cm<_7)=>033JM@3A,*;E ( -L4.::"Uoy/ J?Y=v*㠩oKLƜQ aRrd!tYƬ Iv c%3Dq3rX!lQ "=͈1EJ:0 Xe0 R `),'a t D㎇JE('7E%=ҿ~ RdX0=Zt$tq-J[OBGCXOs۾ I&uDM*`=UrTpfTڢﮇb*4/xQ i";'a1N8*࠱&B"ˉ"aR(_(O&Vuu8&),~2?Xr- F^OIի\CO/y”4"00$4&ԁe<>PISN<& N3"vFR%T3;r}N L 5SigJi% K2/S@#i'LȁMhi2po{03^@=.K-x+A~uX7FlP\L +6|0Aү Z~@ 4k9Mh:4:V`RP^cTƓMm'Ϣ N+<Rʄ&&h6Rz㱴 J'byB-RnqEX!nx3&t%:dq88S1yGc{6fM3C'XV[>S~C;Gn 'ω+'?‘:9qP;iooei˭)nuY_>Z%oMSUK #+\1Jxu[8 EpU*/S rl/fO% IR42wD72"Et/db-VK`/ :Nz@ԓ&=Sͨ%;Iy)y9jAP>b3O]A"{A>3Vΰ=I!LPOgHa<_'ofo'Cf1ᇇýLn- 2h]诓QGn>'GxQqc}+Gچ;-"ٲ7y:M- jfa7w_PLv-4hn\ֻnOnhWncs̀ގK{q@]Jꚅm<}uM>v>.G5-W\ *},{9"9 67=js-f!Vp|1KI׺޹ŕS85 Zhnʝ=!N9Xu!f*&U;E7Gggp)PC?7)Le",o®1h^Jr-} {/(gNޅ3}= ~MR?_ &>!}CH_PN_[>iŒ[I1&^IˉVfT"0#Kj¥J:$IG >%4϶pur>/6 e`>B Tf|oƃ/XP†c~+^k@'6r#";Y`2{8d6id8kˎ${=_QKKl;L.>U,Yob}XS?TDǪ^k9 E@ \4hK=xTBm6>!d XXUɔ+3"(. *Y D:V2RBj,m{^zWJtJ\؛0B(VD6ڬ~`UP jLxI40+ʩ,4QumA ogziGp{n;hxOňyqS)@;_{o`SRciIMX?|r%lF&5N-fkJYدtr[֐wx;UP?T2_,OIU-dJZM@!k y+ N+p?l0r`WOh+"uuCQȁ^%Ƭw?$V# .mX}\E%?E$a.Fw웫+J.՗J (PWfh^܏>^$M?p4$" D ZIsZT1Ѭ8/Q"%1>08 XVPR ^0l%&58  )+wdEA["r<14HBp'pymӀobm>-V;)~4⛟Ti q\ݫ#9??}_ ϟ`{2`#l]|=eԘ<z:D懇EN w߹_V鸩߹[d#hC(8JGLق\s΢Z}q ǃ/7Ke<š@VwI5#$Fǥo `Ώof !wPh1a8%a%2VO=v#v٩/W7w'W9 U>1VI72s=qI>L{3{2srdw=j5;jjb{މFwrxQٳpz(B-J> )˨ %W&xњ"[E5A} YQӖ*dQ~&i5w/hՒyV ^_77Ѵ-rw[ʣż~G6h0÷V*ҧrk7jFq_ z Cգ^\իTj :YBi4Pr}Sh_ܡ#,ȱ8[m@O&s2ю¶P+D㉒O4HAv"Lbܾb^ѝz>20 蝢;mQ7lR͌~Q7wi~yNrbu 黋}l/}/>tì/nhrjrul mC-7X"KㄾwaUz6Okgϔ}f]7/>;s<˞K))} +ɖ$䕋hLU^kKqA21Hw4nDJҟvIj$䕋2 q [ko$s@i{ޤfc4ޓכN9tN_7!)o ]oDžԲpYY(xMB-ξ- f^\zX*tEOsn81-u&ldN1$&dMQ7Π@:ZRH)1Nг -T N+bZ]rRji:4vu-72(+A2jtoTtY#`݅ 10P 1fɭ~y("K  H@AwsDDѭB^O/>\]N]lz񩃅U9|OT^^Jy@WIS#d6zاY Ƃ1!J@vޠ\OJ'`0A N MA) _t-Ykzg9EF=>lQm…Q.)w40:19LoFqFmH+J P7-X [.)&Ftݲ'ڐW.2e3RN(ڑmz`Lq_˱]c{g."NKRt'i'ɉOw9ȅX_|i|xPh'I.jsF{ut =*?-Y~}NNB)0D{v ùQ7a29Rf,G [ ¨-n@ҒgшLH&TU6YTީ)mkOkلI@&̅n3P$&je-N+°hq7"D<׾ɧQ׈$Y1On)#dTlйk=K6\Giu4u}g]Đ~Vo$f97wa*cDsdD%\8NB%=B1rӣ^:PzP))ϩV+aPsf VS^ʌ6F1Nʎ(P 0؞$<̧wqN`;ݺfsb^L[$n{\xabCIrLNnsJ@8-;"LL'i%2{49l@>g*U.ßnI9Yޒ~T ĕy>z,fHÍ} #E,kݹzxqZ\IӫQf{?ݷ=ߌB{^ǫ ~W$)1m'W qyW|FY1vBE׼)vJl ]J82E)Yx~W';Gf) ؗ?uPO^4uh'2aG,y=n %. `'Y⒚t&w󂚝kD_Q+Lu\ڽxm&vJQ#Vic=Sy/?O߶:Rp eZV\0C+h]a! hS[[*yuI4*B04!0v $9H#YB/}Sdvvp}PEvzH#^&g"c&N,35_]* $< vgi8+rec|hJLNSVרSf p"{?軰~"d7````͟&sT-3!en"G p$ &3f1r&PQ{#q>Kįםש/PIZ@םy-Jw| ݁ DXH|ē}9vɾHvy.bkd&f4W0oimn-Khlwfļ d։k:a4M7u0,NըZW QZ}Ӗk,~$;jI3T@)CN(`L;DEIᘥ& Hc8*%NDњLTWBd/bFk\#{Â2f- % r8mp(ra,2[3I}ڮw5t_z1hkRܛY53G= Eoﻡ%-QA߿&}r̘}ژLu{wsEE5p|?v:d|s)Twҹp}"pm*STj, c_)Vj~*2)gIgCpiPm|p^=.p;˩cU!(w\R>#+W΢x>gr(Kz|Y9߇ D])s+LҤεZ`1S\LO(!2\"i-ŽZ WGeOT9Ȟc ^0i/`Ba 3D3al|.wE F {^>dqWɭA"DQYJl~fZ>=|vFK@sRcET᝛ fumܺ=4.`=CXBHPbk+LV*xlb{GlL?,,Fڢ> Eingu:r(e0H{BD$jDUIƒKyU'7ՁVP0ADV0aهg8V9Xb4֡(_$bHuĭt^p;K7>^Cy4i%ڊ9E\#v*9_'JUXpFɁZq^ O%;*`PG%%%UůZ{0R>4 0s2zthoֈRbr%e;Pڃu_wA5X{9)~֞crڵVT3HʋȰuZG5(["\8*^Q{BqI 8:9Sqr)i$&#hYPJyU52Kbѥ&Nȁ)UD(:NQgzߋd-ΌdFnx:!)J bt Kv`%ҫe*CJjY\T2 CHgA1 +Jˡ"5CJZ[Ğ"$Q+%^T2H.F}B]M)c[q@:JH9oi;_p94/xN9*+5qSz$ +(2'}t2@T6gJpLâAvi&qtrW',Lb 0h"8.<yxbaA2ƚ,c"&)Lv)L{Q(hqV89ڵf*yjЕP*_N\|ϖ1ΚIper_Ɗӟq:{ r8"TdVh:ocmyI 9 }zeY]P ~iqrZ5k]8 y](O;_U#7K[$Fok, 1K*uޤ&R?sUH ­em,y*tq:r8!<6-T Rs"aݚUpA! ń+ƆI;1ڼu]bQєAw-8)J- o߲8e?5䂞a$TW%5ƣv[֓X'TvoB~RyO 7iI&d)'Jr~{E (k .X9 aB[nZclpw[ד1._E%Qk}5tt<#,R.k ϤpD%x)N ܅CL*tG Z^\̶R?{2n┢H%hgeh4tѦgx3yatͺ2d(e T޷x{y ](|f:@tW*`oCoQ6QPA/5cBo3MwbYD{^4z,ZW5[U A>bDfeM@j]pe0hBs߅{`Uh:5N?~93)(u#%G`g֬Ԣf+ KLDX -(')5x:Z#Biy-r8 lB؝ %f5;\){yXo$X$PB/xDm! `8!8~Jg=欔JQPY]TJڕ%Tt D}`pj-GpdHpR(I㌠2FѣW? R7WP~RA< kP),+˞@#FkAez"@Y8 ĎvHt#i.omw@5Z˞Pj6=.ÜoIJݿ~KMZrςn&NBx )zSUrL[[nW~`ob9,s]N-d' 2ICe½oLЯ)0%s;h ?iU8m&)b*0k g_IB+HkЖ= mOZh%߲/rOѧdx3gt6fYjyp:vun`?}卯ޙ=L:܂b[+έ/'+#-q w{+6_}t.n&9.Pi [E[Y8sLjCJ#N˒sэczag8e7N }ޚGok%k Q6AHESp8Q$-gC~4N}b'= gAC :|N~Zno.v9N;UfiKfi֫nU[M";~ . Ss th 5岡{ZXp;d@~0H:Kj)ќ@=Hda" !"xܚ0XBn\mOVgցx5鯴kx:-ď]zQ/u1.Fj#17Pi)BD!) )Q2z>G px>K`Q쟦0`Lah 2cXiڃ.JJcA^*K{ pZ* ..2dGD @ДqFGX4((eK D||`#'#T=[ G$hkv A$8/׏UZԢ ~Jj U6\zt@Ñ5Oi<_"ǣ5NOMG`|Ө$\p!K'iRfN! Ig[˼L"Ay2Z ";#IovpD̽A< x'Bs=#S:bΤ1Z! 3OuAx' ΘF-=.lvX#ğhăDT<X55,G@ m$a ;BT1?@#)߇fu*?A[HB* Yj-i! uo2Jv#e$⨼q ~Fᘌ҂+Oc:N;iGm)P0&g|65'I":so`@L6rj#:Fpͤ61:GEP G-}#!,ncYKK؇` cʁwT(w&c$}2ro}??J4xٻ߸q$~Y-dH$r7fgIkl'WnM5%:v8q̖*H}~}x/3HƟ* WN"j#_'?SwO_NZNYn~?Q~|6JqV>zG/J缇6Ply&(´Z H4y㶑^ѥ0Ǫ(u: ZkP]G麯J}Uu_ռA9ʸl8z5e4" ~k'4Z-Ǎ)#" 78 ʻzB !ަXٙR CCR h,_Sn}3%ҤE6gNNęZƳurRIPh @ <+x[JA4jE9AF{^<ѼHQn=28@3HZj [erTӈz $5 LRrM۩K9j$ N酩0]~"s\S"4fx5\/< g/tB㮲79S-ŏa2jY3n _=\nX+aٳFȌ* dj֖ 3ĖX/.^C{?XJgp)8{VFY'GLΨW[BR%&sMלMkdF9Dd=vۦ%4'8(Q cYclЛqJޔQO𒡻|rX-?nQ0P X.`=Bo'cU'莶ܕnSS̷N.[LlIvMPͭ4,\O%̤FL[5b)|~U52hwUY"-aj<_»3FiJ6+>j5x>DP{mŖ2`¨I f>En4h>bDP= Lطi@>8t49r=9K*G`@hxaW* ¬7[I4G'yG|Wk>/Nl=eR \q6E/~'jZj5Zd .odxqq 4m_mmpws]Z*cm_m*A "^`WHC L꠰ @J WO3tL#fe1-< @.FJܞwԒgIt;p珯>OI9O8|F-ӡkxo$0:ZD5'~z8jx0NM8.dKO *ͣ84O]]jQ @dZH:or(Á7;֏WcE =zڍZ:eE{K4+e4$RGC hQtLJdU+PY#9CRd2Px9z&h:!:^b~fΙsYɉ9*9lLd1>\ɉ_a?Y|x7fʴdqPbD1je1bWPP--n';*CK}"wd"dnWN擬\V 0gVk 7gPȚN<Jç> UgUtf˿&˟qp"9Y9s6~W77Ԝ*燤CuĄoҊH085UI'jobG_k| #pRiZ$kM2~_&2MJ+-WP+p:JrX,yq@߲ _~<,m+{g k!~ _>{y˪kv[uCF,HC(N>d*55gh rƋ[tc8lA€s iUf-5$E0RZ0"&R)71JyNPQs!a*4^HA qyN>iӔӸܾyXZ'CNN,1R:"1itF]g>t% /qj^c-qi>˺q"@ЩHaFEd((6uT)ɀ '"Yn`zοoЩi}xο[dQ/ f/ׇWwd‰KԿo+BDxI? >%_+|'򿩸*h(2B 6 թچ2:8q#}4 #͢\DӵEߝ|>|?D:xrҴmx/aE]X'1ZV=gcSLl߯>&)Su 8oِva@us6_ogC^iɝ }X( ^pupe9zrn'c 'y(weLP&?En=ݛ|Wv@ ؃[wg4[~?ș9W9h`_h UV 'gU/*վkb SR\VK<\3v h\eѓ/v27.dºZ3+x g%.:8+/=ɩ k֖0 ~k>JaW*-ё؛\)3. gHM l S܁Fs+eNZa(3^dc[媰tvO—!A1Z=r Dc&Yq(LPsa5vT][kPHF1iLoۃF|(/vNSPk%] s#$bMډ(6 WHJgW4FoGF5gvR/xe$Soڛ<&_PO+'ޠƟrNW|9o88t9x@+1E R|%l32]T>9w=q &Oht~1NQ-:7ZsD'nG nN7Zt)p:iOhiތ NU娥CϏ`Y{nGSl4Ү[㻻nϑ@ .IoktEǟ`VMx^cygg"^/I`xG!Lh672`:Mi!MzPۢ mIv$yxs-݊t#>S[WOk[آtl}ӻ g:T(Mz'S-YX9x}RV(tu-@ךSBKJ͌2:6KG<3q/gJ+=3:pbDg(#G,Q8IJDmdt]QzvͿZoFhlA\aH?2 t(3XِfN?4SyNe*3nKmVY*gP5 Fq2AQ/21hxL**&ʿ7 ]b*yI7t&a!i;;umBMNַ%<8"<>ARp@FN{ xQk-!5}ٜN+q&I$"H2A+CIZ ]"ec| x2dg:dZB)ku0(* wJit LsK& Հ@"Ҥ+rks$Ū Z12a#J瑒gTiʡ=QK'bíHpg&l^-, P>R?' t16ZbZ6F`=q 2 4e|\!a9ɐ H&9#5.u@0Bd8 VKEV{ 21;O5e93b1&\ Mn=7QQ>'X208P. rF~'GA^6PZW UU6h1 RPg=`m dk(Ż Vhis zg`gf0- YipGQX憓MJN\jyovEa*H1EfVئZB4o֖ǙkNz`3_"RNk'] U?ͻ]50a >.ȩ/XnnKj+~Zʭ++-ӚucKtn>},K`0_#~Vz`7&.[<#!Gn9#Eho/k~5\9+gbPwfk~ZM|plQr9~kU`ʋ!r,1%=zpDP*D_wRSV0]~0)$Ia>qǻx9؊A+TJJg;[T>3͖ؕ ?7qE-=z?.K7m1%c)u]/_GXftQB{~[6{aCZP/K\ï;Ti+uº e}Gv:/-괴bm꟰$Z&F|?(fé,o}ûo޾}@>êV6&cl6w䉎i..ڻU311,=,7ud٭ ¯dji@ƣsnm1iGo!X`Nޕ=**%;7~ݯqRK8t(U0UUQ*xeTuFՑV* Uv{bNԕ=** (c@STi?NQ'uFE [=^!C`)vH46 {BRTI;nCQuN]Nڡ=ꢬIu^):= NN_Zңtk'7Ы=ڽ:inKPW Ӟz"}jEKm}pbx>5B} *n7}ȯg, e>;?+=riM7g 6\Ly-a)[\pk о*w ǟno 1-y2Oo~>_|6|,>E0Hﺾ0YFa,9340C)y6w󀕫$"'Ԧh TwZfsZ.'I5x74I:Twܓ+EҬJ~;Zhvګ@:mrhݚb^HUHEKxRUIK*xuАT*/4l#zDп.UP.Ք ^ݧ"QVs3 y](ЌHѻ]h1"ZCBm: $@W{Xmʌ`S# c9h6CF`/Z!&IFA9E2" aGJ%bFhg/ agZn:Yl5fR1þn_vAҍȾ|38>JH3wxK{aBK1vfnՐ/Tv]$c{:Qr=]̽e5<:5fcocoKj0G|qg?{v3֓U*/?] P`uPz^`ZޭguٛsLBI΋LK{7|"O/BD ju+c8 l\\uJ{<<)Ք/2 &3Q7dSxŗ=-/`H蠙.TSKL3U'xB:]||:QCj*#I f)Ҵn7.L%cT BP ~qFpCP@4Y&,Pab ?<8;w7~3fVl? >'FkXX:-\c<w `KfD/+YS"Gy/BGi$Imj% ,N6.Deah !6nA\K)Ƈ{S X1bj >ԅb;ffjys>]3P٦HO(}GfKqfRSQ}`XE1D7B^_K*R h3Uw6v|J_ QADЌ,  <$@eÃHy!3@(+.$hOc$JfBrOG1smt۾$rXOaFr^nr5EX6p2|| AtC`Лzs@Z"R {eHJA2BcJ1˥ǰPWYJ@3l~ ) Hz.3a2ETGb אĺzpiT"H F۴ _6,***b2mkN`ϳ9,u3zPջ.>_ww6e潮跫m#dx2k$' ⍎/ւh7xX )joJ,!ȍD1/bACJ'u {ϕRMJtR&"i Is#r"PD Ré G~=Tlfpe:=)I {EALxne4KpH_ -v5yqd9bn'bxW%L6K"y3<ώP0w,S+k(q3PtRqIĴcCkAﱒ dl;W!B \5 1Rj\VNUVP:#EHHHFRr՘;j$kV *)8 XvSH0@RFAF)B Sb{AFGHMZ%L$31։Oq:UYJKuUs7ijW AƋjdSi6XeSOC_2pH{~޾Ci!גA~j*}DZ]Tm#}Fow-ϝ](}??Wpm)4*M{_7?x4x?Y?{Ƒ ={vGku<`qĈ Zܕ%HIzˈ=%k˧ⷫ^௓ŗWX t`+V?q&1vޢ2,Gú6n\ /J'ULvjw)[dn◩0*Gńd\oķ>g.]]= :n +7ɤ| z4UU -U:@7>ߎߞwUYҟfwgso(z0=hV(l47-8?*Q,Q!!QعJ%]{D_\5+ˊbܢǹ=qkp݋jLt(Raᛜ_v+uǮ-}5}ˤђ :Fp|9=)F#vֶUhfi5rjCS:osy Ax}*AtbbiijGj ]ު:\jʈ2Znx7Ym43]~Wę9v9Lrk_BڝbѯH!eP^< IM-{$Wu@jGΠ M$ "}k' fͯozk )has:`ELsIwX( Qd IrokI/aAn<7I',Ac a7J{55^ ḘlTdƃ `-j7汍*$Pj#_A.gwhƒ2.1]Bh"Q)|TRTB臙pӚ(nZȔDc rWfQ`rw. I H#sOQI~ʊ+ۄ:PL3zSZ32=XK\Lq]~JONpkSY UnƖ8E*WBnoA 0a[ZVLss6&uܷgkm ։SC`N7Cr:ApjťȢE>ԩq`$q1hACh)TƨOQ )KA2^hnRamm[)b;lPS5BP]}u˲܉HЍ${:3G^D^ӛ˚PŶ-MhQPޚ=bt\ˀW={kgmkʶcC']^&#FNR ^!gY{$#02Wt'*FLAa`h'`a9?4crR *ULPR Cl s/s'4^R-pLLpFHL[p&!. d2ahm8*A*! bt%띞;̚1bCChE(۝uZŸə$U &2* c8&sk_2㱑CEȇI۪ Q#\il>`̜$Ye]e|FYeKfW;mځ~غmL#@$ݾH9 5 39nʫM< urC[g/̴F!ÅtWR)E7X S?:MjC&d\=QHpe-ŝ!PHT^HZ9Í|Ć,g HXdYte.370wiư}ֳ0f6&cf/1Wj_,묐2<ׯ.bL&ˁ͠!דǺ[Z\\o {j S7m6܍`*| ({7}d|-xs94{+f=З ުhoc>ZF} w~Οɉ΋ⵜP ,E,43g$~g~"%ZI3ө?V.1h?>'ǹqsr7\/ENF'&K1<8 ള(8Pd$9 J>kߒ?N_^Y@|'Nvgws;Ms̯pIgOz$6&:LOc.Y)-P7_LBŐ;lq/peCLqvRa>_|KTg=j^ߝog=4JFv J(>dNUi.H)Wk :\8 p.x l*{449A]Xqw1&( oA')%8d;Q/ew'Kd"ySrwkɸkŊ > Lnɻ\A_x:`R6K X_AC9OrB- FqF&fK8u0$zG; __ˬ|@]/,Ir3)B[Bp7$3FGn@IkIB25-AL $ΤBaVqs`$RPZuI fl xR%X'p]C#ypVSIo, wɅDCn +sH r*jcT+`Jo侎OeH~&e'Žqi ރ^{8~ 4ǵ|^]g,oKㅗ-翡'jqvk]o85Yg|~x`Ve?q%Z*wo 1?puz$৴9ߦgIwί -rݷ'·e֚,5 W gJiA^)ߜi3|}ϴgMH pFy!%tI(ޫ}8tW@\E4~E$pV.T6c*Iqt1K*0oǕ`ԻloY`JDгe8!2ipU'뤒P _8%I4FUT1 w)QZ{D&)q%%w@U-5 o~XIإM'T uʰDe MЎd^r;K[*[r)2ɫ]0*KK74c&kmdNPs0Mr`q#E/w]x?R){UWn0$*$ع~ % 0áˡy4~h4ྠJa LRmC9Hc"e I\-CQ0E=s_|wK6L1"N!Q +Z +# 1vDdfl=P5!ōչ*KKKd/Kz ~ y H( .in-E`B̧^{]5\/.?_0tr , Bvդ, PX"NPk+{G+lu\"p$e7^U$s!Rহ\S;V6 O8鐻2e v5qOP٦$NJe FoPBj~3?‰Wb#Cf4݃ҜW%V-[q}$Xx% ;1jȕD].nǬ.(:)83{=㶽S tp9>wVk.ﷷO"T^~_{2 >{(]]xDiz riZ}Y\WJGNA}{)m.mDG< f1?8<'T_ɦσsi#0 cȘiEu2e31oƼ7? |czl Bp0 G~^$< >=zvp;|g:0bhN%İ#-3 %dj auˀ'a<; DH6!h vy"Z}:LJD vfrc( mG"qrOQJ﨔qzע(ŤǸe2\fV1lz K[9.=[lU!0Y=`dŻT4/|2nnEϹòV 8tda,1hBp,;֜q1&m[;Nz.o-4 BXhU99/[Z :&3$˄8H1.4d&1z,{;tp-:z2O!3_i YV}!yfߞ:M?9O v[Cuwf̦pˏKfcX HLD 'BV[Jvf6Cy2oA<{'#hɁ8).pʇV{RmkFy$@plVķ5HHS|=:h6-z6`RgF͞, ),<ʷ?VxAO!tb" $qi~)UMEQ"uaR)мPFb&*RiGH}uj]&/N~H<.]>+g]-wý?,R>y5د`Q?v𾛻^D _wW/`3/V62nX,46V{+~=mA k懿`9AXP5fa/ {hǧHM%dL>YzLeOE{be?i%qz[AD)/%('MZTI[{<Ȃ~ ,+`DqmR+ImCԙ⥓s'hC jxLB HϤ^g4|jao\"` ̫^է 4A̫T0/hĠ%2`Y ,ndM^!ݬ$E>S` j5/)IO2JMʔ0Ìآ*=E 2*m8/+ _qV$BFR60cuIiBt4S8h66+oRMiCHyq??o˄GuY_kd |_271/R߱ x9uW&B::RoZ 3%#7n%W% Ԣ'[4XY do\DdJ**vlD$(_M䑍tTlѦ5˙s8!:Q8%@`@ 3u!Ȃn;J& kh(z[2u0apW 'tՀ<|蔼. r%9B 3E9ѕCxj՘=/zpR1` fW%̰̇4{NG/U#??8-P%57y}f"-IYL|+N0J}^&>! $9~#j ^v0#,Eϙ@aI%eZb8њ  ˈl! w<{3vvLhz5 A(l!3#/w/CZ~FlUApds}*9,WitqhdsMFJ.WMScuNLXij$x\ХxHҥ8ZKͲ jcJ(BаkW(=^+% ?j#9%$2!9Oc J7"c-Ӿldྤ%=-C%2A&BȪZ*AI(7pc0DK88##rTFhL5AD5JTSKR1O}GLb1%kKWĒ2%d,CJr*gzEoԼ֬z@Qsܦq4c8cӰ `1D,(q[5;ZkEP+dՒi]HLzk<%%XIay9aZ<,X!*9ϐ_Z0'ݥ49xQL_4aRh̝gDW"OdശpDxtOrr?MU1_L@%k(竕e̳~:gRtє4SV ѻ "DQ%&K>I q߮}-fQց o'bjklEeUb^V\U2")M5Ғ2 T !V*m u{sn0yrXܕڊTLWYjK^P 5IECS#E!;11Yz:/`ɘ^N#Hf1|ļpТR0s@8FM9L5LbNQ-NWՕ]h;ud(ʀ 4(ɉe9.Y*QJЛ VD\6#wE  tG1 '` zި%KY"gr$k`(9궳hrh[H@f \Y6ӂ c]j܂t-;[n5Za5r֘(ѰmX} ,Y qHG3I{ΪyR`[-50X֚l'9!HZmy(#&DÅ!mF2RQSZ&LhJH͘}* ^1v1d宫R&7=AeaYK P[LTk-5BtA0bp鏈IM*+l֘B1U*XUKh!ZPҬ̐'ÌD "%ϙ9MћQeΜFBr& 6M̓f(=f\O/ҘQJS$*8PXcs~$NqF"7V*,xvVSapHuRr")@)Ȥ0fʉ$Sc]p53ڝ (l8N<#!|bG`Q3c=aʥ+uTh+ؕ*6_$= j=+zD2.-a7Vd4FUL7>Ҕ\QyA5(Ɏ棒sl$RwmIWbK䭁~Xx{~`yXB^-adICRmIRRI$,VVR֥Xy"222*#0˻eReӿe!5Ԓ h#' #L3q91އDXb-7Q+|F#mpVLp~tW>{W] weǯ}o#FZ}jZ}PjZZz cW2A CVLV@M-hJCn mQQݞK.z3`/Y]=QF0 aN]ntpEnQr2")X`z+zweiczWS_\W>!:$N~*ZP&ĈVʣoO" ?*)\sW2vp5p!TCߗ[-+BY|OH^jtL(Q͑WD:܏HzMLFeՖ*;b" r ycQ$+N'hGz="{';I"^٘ WżN榢1!8V0)ƃ0#jZV+=7I'h 蘒0*sGx%56rc*2%trOvR@8ȓl=ƽ4&9Q)'PP ^ Z.6 mz61֦eI2M;<\ТMۦK'˿3˿lmų3v|7)GCxv;㖛*%p!bz|OK Q^} a C(pqC š-R/AR:׷?/J,5Yv8K7.\/ng9dǍç])ĸ{_U,r$1Kc(Z>c%y|p`Mܔupy%}"䩲J>/UG]cTKE>?%F أ@_Yʔrs\4T>_B޸v)GPlz7=ѻuŠt>w;⯉$f0=QjX76 mJObUXGv4XaE9%Mm$w0x)Q50^wѝ3 Yg~?*c,),13PR8St֋8Y<ޔ@3\YFm rA}HpFG.@$yv и_hD㲎vJŠ4's%sc2pᯄL5hЗ^&üxXMZSY3'?|Wfm96 fSE.`rS1l&p-F0:"zu{{['Eo%6q%oqƝ`UlEw LƕkmC6@Ƥ*5i*d7}^ }gDP]v|r,Uk 8Aϐ^2Yy4r:!Zڬ)DLִ8X[l(iZ??Ŕ ;HK%LH:E΄ZqIUP\@,2f>HF2%@ P1sDh543ʭ:^IZ APR왂 kx :p = 0K*\̇z*\dX"F ?i;ɥ5qI| pU"=2 S-eVq0؁DT伦 Zm^aPQ p覓a:9X(I_IO'4aoQE`.ҕ_n*LC6ׁ}Bj[b+Tu}<)J>}ЄXr *CفY1a) e2E2Rvh5; h1x9UYX $(hK{V?(O}|~tq`9QMe va}׆|kSdZ(2匩a |P R&mڧg'UBE/qĘTRb1$`?!u;$!'18|t?7  @f-Vcڎ-ΓIլu-NyC)˝0y LGI!Z->.B+*!ʴ2ʈ8)i!*y\4% (1$xנdthA!Z0#Nm&RGA9Fe B2˩0jgm>:&kZ@_ ZȨh3聁幎qЩpȞ1a9XRZZ`1{9X6>c2$Dmr fIXgټ6^OO QjtX_oCFߤۿ5?/g+8b |7^0z @)4 ;m΅r{3ChP IAb mTԑ"Tr:e=w)n0x\|5s,:Tvǜ_]ݬm t?Q}3$]$nԅ]Dp(,qq+b@\d CVEupňKP6e* ~"(; TTYQIkd&M^Hw_L&fWPƯ ܼᵯgg"\'aZQ3kՠ Y)z|p8Vp J'NZ'E7lrq \Cqv.V'2i_΢;xpMq6/p1lU0[e׸7b0>;sDp J8wPv_)LGƈ-'XmB3YВ4-#`A'nC;1bȩ$0>B7U=HkHѦulm׼BycĖg׈g k _᜝עm WюsA+1ۅb~k-KmmR~ՕU[iTJ^FWWqry]up \nm H[mm5<`l\sW36*D;*pwl*}_-G'4+8R!^QMdМnr<^1p`gEvSՊbG4ү哦h=SXY4@ؿĞ7ȐynZEk ~E',Z+dopo#; jF [Y 'BB Ds [ - I鷽! Ui)bo+F?&߉]ݗcQGod/(ᵽ{]='9H7U!GM!S$gE jk?^o/$D6R 2JOK" >+sb㟟?Gr`ɠ&E y\ rќ دa1n$"35 1LT`:B"MHga Gdzϳ9'<+I=g?, πBOçaZ\CZ\]?܂EgKE $cb0%B7:r 9eovԸXԗﮣ&zwv;\~}³ֹe}]\~:>\_GVRԎ&1RUϾ6;tPs~Sƹ)"0&(#NI!"^,so_\0eJ4j%Xͭr>(DrhK'r)eb:c\7S,x儌)6"lT="D擵g&Fz')gʚ6_aeC0!ɎY)$bGGf l󐬙, qT$PH|U,LIBL)7q(8vT2v1K&m"+&~ҳm4.䪎׻XVSMڳvR$>[HXK޾Y}]0Kx=W 0<Ms#ONg%/F#Oo-ҟpOgdf] 3Ya1"I\ݽ1Qh (&qي]AW8 Tm rMM~dž]ۑ E/dyaӣecѠJ>]^R*=;E_"׌2CDAنЦ(TwB$ ô*dʒdB db5?qكߑuN:_mtC"8=}L9j} ?n1ry26fl<bKl ǑE6Ɩ%E_M.!# 'mg_<_ǝ'\:hEmll/ȶ ̉ՠFcOMړۓ}›۞.e&VM\\7ąu?jWjL5y QpusR;*BuW<֍:]=xfG_yՙu=f0DNz;xşS\ DBy>U@%CS.] ~wy(;{ I1%>6 Ü e!'sBb&iDbXQ'KZ*˧ph)HCh܏ 2,4EbyވVp4@Qd<PF[ pibl)g?oQzAPY9!\umK++䶲Yo'%_4G҉#Q{r6,QL1nB |I"KPG<#8|wkT(;Lo451 uzi E&(Siiqs0PT+ݍ&nnÔc#GrV^(XXU JG*I-]A%i.UާorKQ`5ҀMi!D`4։Ҡn,w1d#b41RSiLTA&* -v/?"l^v8l^昌r@(߅y`LF1(Mk(su??Ȃ˸eRNdkDcӜ1GBqEFXEpONd9dg";zitu O`db1I?sDDpǞ%ԉ@x2&C8R)0&:aU .qk[6cu k`9"d|L7߇{I m&'%M>-F~h$y?.k?\:x1,h5}E{MTyH௔ 7\hlԎNF.1V 2 M>e?攂97n%5N+ЗBu=d11Lq#Jy0 ]!_ 9ɛɗqqnMxӝ=㣷c6flI1õ z?;Ahi8>~_hWVᖢBu[k1ɭskpҠxbSbKޜ`S:jF` Jfvlr2]8~~gS6Tb*g'xR&NCyh{3~WWV|@H',vՃEkNMsm,+ukT fs=ݳo -H8ac%#Hepݍje o9D0FvI;]ŭB*z#$;ȭlܾ}&n6 ,6GwT~?\4ݒQ\sb (VPl >$%%xF@+ɣ\PIyw]j$mTqD .4~i(PT4nQrqq+0Zǹ?M'OY JO ?l | xÌ~l64~H mfh̺ÌP{ 3t.8fIZpVbLnDƜ] $PTP>3*P4;;VA+0N9ۤz*Z@†'8 |2嬀f/UՅIYޭyNry&g~n>&ϛ2L1>kդH~ii/w)Z'i8,p^3OFvr?czK$TM?ĽD)1HzV2k^+8E(ڔP20'OK)ZgnCf}©M?9?,F6빶9[TJM6H?Bn ?yfIUh0t<׼~ ;gV%+~>IOx ʬ,eMEL -]nU1(#:Uqv; 4WiVoIVp$Sʣ['SS" m5_X! n)qd"]9i]d,ЁJ|_t| :γ} VK\\i6@,a lj^[f^%Wmְ2cv@cEp=܏),NB,'R.6t(Rvmush?.cV>`;i&hL3 JJT.b w~~-E)ԧGN5RcTK%).Z !Hç3* Lcͬ*_V঩]TZ8U# |V?-QC@_7*6 릤JnA<`aKoᘡ)p\(D\qyH7r;9NԾRNkus3JJ(FZ3Y3/ܮqBT8Bb%r Պx|\sIi"ͱ2#FH$F>I9 Vң֬ţS9|wp{КW; 8ݽ^)wzzPu.E:;3?[Ҁ%K޾Y}]0Kx=W? G!-W ~u h8tR( ?{Ff6aY=%21\, `P-1{RIDb]A8 R M2AO2ơ!l1iM䵶˨ Fd`HĴ&b=Hn#m'NVa3Ecu3 E0{ ; Ċb8_|O6BaLd2Wv{U~rQh/ . Hńcm14ט; *kmH /]J~,w{18eFE$$~39$gĞyIĒ؜Uuuuu=4+ #a'<-~Azlk;Q82-盟~Uk).¤M2ژLd9GE' ]o ybR/+e̿fQ!Q;Kp(\e!ce6gʅ= OJO>jWw}}y* ,4BLh)4)ʀWV[?d1n2$/?|j򙽁5\ Pǡlס1Bb_94z2 3bGrp,3HC2C-#O '>EiZ)ompS`/ހdqjBT~-R1l/<GޣWcu@BR$w N@Π8vO A HP, pz(a P$_%H<1kK^;# 0b(G4,aPWA~º}ND%. s"ynvtJL#VѩC*+=|9)\^\ޝo=L [=|ċ*xQ.^]1WwM8|`t|X1zpVQ|X#:Қ;{8:8#+hzQhZ,]~ˮ7Λ:ې]H̘@} 4{29”c aH[yyfXd2$Z:$3eg&su<]F=SLZ9XQ-T k4ՒaT 1gp,T%YQҍ@w˂up:TܑF \WRkh/SfÎ2ۘc@?OPҙ/ѿzUtbR5'}-'O) wXkKryzhߞީ?!ψT⅟π^'˸ oG ϗhM4 A8LKLv↟d4Cc1|H+&-áYZύ[EPG\ c@z;& 9"ʴ:U# (₃SZ`Vlo},h79@MC1~Ip:c XX.鵽Nz-Fj2:JzmQbZ믃!Wc/IX(Sba[Ǻ39O_/vw/f^31%31nx lzN(sLikaNH+$s̨Op5Kn #mWHzqdA%QZAI]OM HpUyնunA]8xkq82$Xe8$FsJDxXF=xp ^1cm΁ BI'.V)[a43ҮM2)TRBc4C&wdHeHiiUC&e\kV)2jj&IKc;A݄v"0p8AI4v/bsbߒbieԆ@dJx_}~y m^|Y^{-< SVӈ/eT0.hbi|!"/_ *dEG R3iǐ'Q>5$FL7-fR 9 'M;$Cw%JKS|?=L'{0Lc0&R#j58Ŷ:IAu6 Z zd@;b?;*qwe ׽SZJGU`׃xno\Q"Xo$zAhh S`*g6wa>gb͋sM5/fMEˎ$w.)28 DkT bD'MxT?o^2Ru!!sMRF->j`LvV1 QkDnCËy`nt0MWG|96X?w!(fܸR]&d,.oZoPR߈|j}@K\&ֆ0#H_ 3hRd͸SOe*iKT;I#DT=ݘQ5YH7O}3dPx)w6>ga4n*1LJ䢯 N# 7=8ۅEL IXv,>u]NewlZ!W=JJU↟Wi4AlBk ,\3Il%H(\˞NP`%Čտӱyǔ!g=VmW ӏ>}QXoJ Mi g DGZdzusͦNIgjsAsE#mH 476ifr0{ɋN+zd6GT8߶LAʆ7bC`rll;Y2b[5Y3 awҙD<~tG p@ߧcX>=}>AwTZ'_(GF'!Wכm!5xlYQQnP& @oְrZjeEl-FnWUeDz㴨:#ڀP§V5[dZ5[pKĭӋZU w4 By=12d;``a9;W3Jo=3?S#, mGV"/W' ik_Gg /^:%̆<]dvϮژ%/>ͧj wo@u1dЁpy%!dWLd'& : 0z#8Rڱ)K(~3=W&4y jVn,?2)A!c_@AqzE zF4}wR@ʂ\I|,Ub}$gژM~ ^Y.! b}-ݯsio?sA>y\jqwKGd^@ 5r1h!X4g)Duj0ykP6 / pPh׳m a "|u6YR Jmz/Ă")C倥 Fk5gjd1ȤP"0[˰*@aȍ @$Iڞt;N%=أјK譊\qMDd.0^m#:ATft>Vz7RD$)$#f [hN,as#'-X>G ]L^V[M|\S&S8{JzS24܀gd`+ك@/G8rܬy1ܗ`*1Z3\XP)~VF$PN&j_8:݂}XiGlǑ @1xKD1˜|jDLgR(H\sXaGjI FHeHCJSJc,SI4Q#8$O 0̐'Qk<FA,X:8N$$#'@` %y4r`Q eDVA bP"mAZ$RB'2d@c 5w4DkĤ6D:ʅqFq'P䌐fFaA_sGIhi)IT.zshY_?oVm(<MIӡ=w6}~s>w_zrB*b{J#ޮ\k_t}F:/wɳ>,[֛-/o^.w߹Rͷb5;7[x[.@`m{!8~=~'aٻm,WT~T~v3J̼d nMdKdw2S PEIIn;lssAXqoAG_l=ly+Ïv4[o?-֋'z͗zA~+8w]''K˓!>Ϧ,;{=!0Q_4o4ꛆ/13$$]~ew H [ALl֋^GׯqϏ}Zy!~&@Wgt~%ɏKՐMJ 3g4PBz1|?nvWf-G49Ƨ&$LKn`}50>xN;7'RH)x Lf):Sz9MN>Z+saHkpd[2șQ:ZjWKÀlFC>w . H5rq7:C-2_FgbRa;ޘPv Xti2(Ƣ)gle"24aV@'@‡dj0@B"RifR@c)ys&Y9Y d'b64iiyړ]gd^287Fhs,FLyfZo0{-F[=lkd|M@6nJlGTfCZ͒Ϳr_Q- If "kzV-O=M_=d &ۗdP ғ-ҏܶH3-҃>"Hs-_a^B^N<5yA-h.M _ XG  زNy:}˟%-'1dqDȷO y_:0kBC5jo }90P n$R'{s~סL NIh:O9̦{j=Pqmxui^~Z|aKDUm'+s8G+*-3l80F'Ֆ?('bӋœܕ>9RrHDR%`3/MNQX|s޶UTTU˰Zn"l(\6kcd;hlaId^ Zk6FxAٕ 3VjXcI)82H@j̨ Ҍp}Ҵ!^">6ԐkhJ5NߠG;PPQuYG'm '&*85+PHBcAMfTp+ع;ʍ,.5<)@gq<`iQ EufBCnJҁX2BaHM, .H*uo $սsQ| wg8Zj6;>iXGq4F&vLYzI廛Bv}7|'oiz43M1> Lun76]忲NWUӂEKJxOz3rO[N1ݲ~Z}9c'87wlgJޛl*7-iW:@uk aukA4Fխ?qZ(5hֆq\%BZ{ZLJŏw/,]w*kg1񀐌  FXȱ^<}zPn%܏*7{1ڪa֪`vQ6&eV^c8体r>"ٽ 6 . ?_VdSRʎ|  c1 P@fn#XHLu$X<"mc~9~Lv"ܝJb7@!*.E{\&8ZkY(GO.FGGhKBb-9!:RC'{#!>b͐>u VY V@r +lc}o}}!x䧛PI+ 1j:w;NEPݍOyB;w!o\ET~P7 cB5 Du~֣A1Ťխ@CW64䍫:E#l78Vqo~IkfpEC@0f7K !'WҍonLZ3.F^.`h\_a @ -LWddJpTq@x&Hq+x|պ{`=*$AV fqau/͇Arijy9+|_oƜq%e)hqd.݋Bø8]CJxzڃ[(] A@Pz2ݑqQr] "cG]4P;#o2Z/7l;dYWrQU8^oȍ XGy)b!?:ȧ@^pFo8hV A$G?,F}8 (o1Iq}_KDbvN0%G'Bqxw6U|Ub2_.>8-wj0zgݳZͦw>y%(n&τa1(e J$uSHlf2> - C)-cX(xk:mԔINbT?QM^C,"::M?dkTI5BAyi !"ETJM,̈TTaTL` %T$$x#XLcn~6˖x$b]ͦھOK7tfOoO~~sstlR N Np4&ͰpcYbk8cAMG1~?Gš L~~-F9emA{XDjj@;h[@1YzP_gbBT'H|%ZzNכ׿',B=4)[,.G9xIK* ?ӆ+I416!2&Zf@STr3 tS'NEw@a~u_lŭ;_vx-+'"vO})Bt,MVA$Ƶ$D?cLsxa<1"bEewI?zYˇQcG uģw rsFBNXnRzRkOFN%Uz}q mѺNW_;t[uC`lL+JNSawm#~Ih2ffvK} +KDb6~I٦n){gc[4Y*a,#kq" 卓D牾J 23nQ+*ֶo:χqxZcxT yJ0X{xdj"5C<`$R/< e kLGv5IK8 6!fÞ8z.`v{IDM蜣mKL,VωQ;u> 4+(xW h7-:8NV$?<|~U`Y"m|+EpVe |bi5_nS31;5T?S> O6C:jeXTSSaG27}qMG gUv XU]i&[4j4rn\(ˉܥm+˕kj|K)y&_-hwy>wsˮk;% fߥ6~l&GWaV_I2Xf13 Ȣr^oًSzeE$y{@9<%Z/`rB>4HS .; !y-aev9T \sՁdM?b6QX:"Wt;Y/mIn??T>-w\LKff0nj#ے?eJ(6>8Xs FcX8c;o-'bV: 1/28BEX3nI- F(WT-F'-8}Y -ZlGEbDkAXT(3шba)FPm6?p !PsT/K\ =^C2^29_<@:/KI^._9/3PӁg|E glKDL%IX):<apI?iD3fZ\ .M0m^ܭ4\&=hj֘g׼U39 D3у{u>𯞏 "uFA `r^ݺ@!Fx٫zTGzsz۝ xݐ2vүg"7@; u-Ď[c)ԐA5F;3U^{xm{8Y~ A1Jz}%agOܣW]~t1,YDoxAƄe /2jCn nXop7]kBx{1HP&ۨ:Wك՟Fgn&+t̔:v:g=|(JS,+?ͭ8!osoGn[βԲ)H֖}昰G[DHAVO6CTh5`4痦KK%ړ!į_f/'NfgظmHpVslY>d`?[|vw~G")2 䜢A{ϛ^ %-d/m4#Q D9XĤ'"Daqm1TIܰs 9!_E yet1ϾGԻ>'?]jEc;^eqDZd\"I.c(gcX/0JSpT31>V5kay& #ͭ^dx€+?[Мœξ}eE3e)i֤ \vRۤ!*շUA,p|$zx V$lh5H&i}-5W|Dm6z}"\o>uq]"p)FGDi9ySSIJ=מ]谿.)Ҕ LXFJӡ 6$3^H'S^C1H|@zjz?a7ҥBsq.0qӣk/یMxlKD.Ʋ?qϴW809?ft!Cw{^))mSS,&6Togh r{sw!!rD0 _4wXr`O JiRsJ1bc.: 2cU@9X XZ̄ ln!5кH)P}RXPzvq[j^-0})&ERrњO7J' bq{EjRHq;,.yczwC9W uyޢC˯AgzngaYBe>_$9LEQ>r> f|HBY $lQL8zJSJ3WTY];WnlCnm1c:mĻ]0fr|6Ez.,䕛Ma#H\M[Mr@B5&ρ Ll}◐'ƅEw&#b\oIqKRPe[W|):FS8I_`Pl$z43JOy )L|t1_a=,+.j:KX8x(*x%wqЕ*`]BVvP.5*Z$E%C[1Orijŷ*=?_0F^[%]+onU=DDBx,'uqX; [M!E$#Tڌ2mϤC6W2h} cU/s.E[XfS >v/z-xƋWS3g_)ystq7M4҄Һ'M&״~oSǩZራyf~ov箨o^6Wr/gVB&˗kxنZǚN[- ;-֊3mS5{mp{Fb⥥Lf`;4&pǘ)I T(4#Lǘ*He1c) #u;`Z󤖪kBRnHXʸ 4cF fD+%ʃK0 i+!RV(9'(:%({f6"mB-^ku2 &ذk\axlmP0un 9.8("Ry8S#Zml )=#Y4 ԜY#e<xd5 xAi2( Ih`ň1n9*h06 n9D{v.L&ૅQ6[ f\aF/YVDE'L(=` ,JD Lȩ""5 V{v\-eOJVH,ҹ5:MDh#! O7&:#P޴0'pUޤϒGj)%'~ _zEz׈u҂.HmV"XPtzץj@)&7=rh(g#6# Y CTh`3d6%\.<<禔K R-LFP)#$66<,@%z7PDMM3yⷧ8rUS(_ 7|ߍn71W0*X8.nnpJ? 4us$KLi?6Q+y\Xhjbbui8}j~\ڱ$+%m⼔J8Sv,٣5IoSC]'Y cԱ*ߢNU]'9 o23-I{RV9[NvƯm\sN᤮OQRh@v5KM9ºR?ʗB: pnyxF**J'{^N ÷ ٻFn$Wz!+dމ=ϓ'( (xI`I- /Hy;[Ozadc*7\g7#S. DC,z$N2V[8Rh@vBwKn1mBÿ O~wԸ`)(N+x̨)S&R]M.'OBsbE6&Bl8ҍ o|60JfCf;PSqSyhO!; H.tКMS4hko縕] U`GArɦxD;V%?58Zsjv2FJ1hhPE@SB4E)Ӝ \K";Qp(ˈ])R;Q;deҸ(QDlZnhWLH$?\SE:iyP"bA)zOz1ߙ:b-[m\e\MITFl`j-"y+5WijglG)ZOGnq0ՊAa&Miy::=K#b|^W_??ܙFKYg24ыg{s#Dze~h-Tr/_v-] ,?hлpwEn1e!bJ9_}G&~IZzujac<.A冀)_4o:Ґ/\EcdkuJrDV9%x~WT6.neZ.4 WV:%`uci=uk+iv]IJBP֭}Lօ|*S:Stc,nidR.Is{v3{:1Z Os0ʌw1P G[ϧcqJG3^X>Gj&4m8NYb8ΣnJzoB>*)>G7 ,w:mRjIlv􇻧ܻܵaGhyg-A9dW9כ&'}' LYD4O媷L2Ԏ80ՈMWx+T.wtg_};njr$!QcC# - !Llâ!ǂeCPt~@Pga$B@RxKOJ1l;F!zGޢĔ؁K"fUۜD_^51jW+Ea_(G*mPEˤܼ bOU\?OCX_Zu̦9lQCь;bRG[J/2tCȱ{0Ҏ. wAFFo #tVr'{MΓ"MQ>f‡ٱ[fh ETϽNKVj,56F8Oّya "h i>!H&P5gLzR$:= T.zRʬndHsoNLG I]@lZ%Xp#0ZW5h YUT5yziQ*ݭK 6[P,zyc Aڕ)e0kQCKS\Kz1wv3>#='wɝ臬,!(gpnrgMFWɝrĿ`8' b͖U`XWAHnXa.R%GꀤA^B;pmzs|[G*[k0MNE9UGilnߗE9~O[:3$4?4!>6();0dܱFI #fT#|iz$<#Ιm"TC!X_\'4ѐ:)>.e({B洔d!w~{P(@8Y7(?G5p9=  Ph0 9dM`|r JewGaYzbΡWp)_(\QRAB̧{qsK=Rz*WWt^!יV 2Ʉ)h„8 2+4`/pAР;<S P$f׻榡"Szb~̯.@@,6:@W1](ٴ<-xXnrGfJ x3{d^oҫ& Q-E/,D ' s Rn]: /WR"F"ΎWˍWz 3(?c˂h;\\ [kոR\f~&,b0T^b#Kl5/>N =٥"L@KUmM*q{Rez@$`\z,X#H ̠ hIj$ r'+[_Jj@<)e%wy%י}(>O7R"f*N'ND!<Ԃ'/Sm>ɔhi뚋:KMq'E2%lɎďv)t&F1u۰dQQlr->bs NW,q3P0^PY<ظXUR=$wFn3Uų5_fi+am[`&f;1{$qK͖{Kl)?ݙ?\Fb%٥D{Z9j?,y/ɆWK3ӉiJ' !&!mByebоںYr2N( lOR =Skcyܻ1,AưkWD::QB㝾]YoI+_-:`z1=^yQodQJl%822"#29B4 FVGJU}JQJ2I ^$Se$Hcq46 \.1$R", ֖I&飍_opf>'8|>G] )w-./f3 p>_ETt/]{QV_6\i ŧ̧ǔrV囖oR:WdՕ^vo6ƙ;χKlX@wF-}Z* TL!ZlSg󿞏t\T5!F.OpgNjrn-p,pa:od_=jƄ OHrY#B)@qT9l5%!'~W 6r8*Rl´М+%kp":ds: 'If:QiXifO2zB**άׂ ⯞.[e#swFY7D*_'Lxp`teM`f``}ێYzzqF]w|ͻEEok_])߽L{u&=VO\a CD5oލ~om1oKqgqO젫~Q-#eod} !Fdmܵbe70FKnIz=+`% MlgoA͕k@(I;fzwqf)^;}@ f8X.wG~Ï<:\.a'oHΟftj5$Jv.vw/Aȅ>{Movs^1ąǟ4.^\7RMA _K}EwT;=xF-gvͫן½{ɿ{ww 'J[$<7 J(R 30p*D.먁5.yzp<>yւcOwwjW+j$?a%;nsڠwe '`9xX[Pbd59ήFVrً#@7SdV*51>A6+)j%ۙZ岕>* s][+&ը`y-.^=}pU c6L sbFg8Rv I"8?b9n5y4 GcՌܬܹV P 4|{K w~ "߀B v|>qQwmϻTJ~x%S< ^ȓbOPƕ G,z˃hSr)T西TX2|*K}AOu~b4ʎhDuu;2m8uߎк !߹)8p&9f}yP:}ƺ1JÀYpZ!4;W5* SJQ6pr, a¤$ Nڔ5PҽyW RL$k'IJ/FR'RBK)t$%2p1: G(捋]m:SAjg HB "$Ð;u'zA^m Y?ךE؅|uKbEuA(*Vұ}2I6\ʹ\,uĔlۦeFap<2@}0i$$Qoi%P8HhWgnײd-e+dh5A WpTVFRZq 9X2FI*7}_/us:&f˫bd * lH1 L #@ {zmzϩ*s!VFJf+>`EyaZ2BT@hHJ*+c,:CC=ėx^H<t8W5ϧ=T 7FTcN^_/C 41;"hF1$ԇD奰f* qzkHLv%hmP$ib<#v`NVdmmY0Xkbb]U!Φ7_j9?bf*4[+[(8* 9*p 0B3F8R K|G_CmLX{ (aЂxMw7GncYuf{s{=Wh2vpi֙.P3}3/Aaɒ1%_;,ٚhh?|# Ӹ󋈋s#rTa(i9eK>ξːn z|JGf( /,D=@N4NGhd;ruKu3A+QIݱl\z-OKVY>Vf`1~6Fe?+.ljcenl~9D]F,c`[8EQm)Hfw_E}]5A/ RE37Sʟq 2$@2+ZmR)I.+55%`Ꚛ,it4[\7TA}wRNi<% ߧ%<*,`7 B_og*F닩Kd{ً]=45[ Ï a|?\2FiNFT/K߱ĩ%,yh)j.x&,5!mu &~$<.+ E}?]cpQS5'pRK8eQQ2IB"M)P/eiOZpV1M ۍ|5Brϯ Pޚ K $) :(M% _h;ٻ>]w 7\j6*Ɛ}77-o_a/iUlv=#BEyi!Ob߅|"BzFʴН R2!U614R#`-or]zust˅_ m). p-VQGq^iS(zꔌEux$IyF}Zeź?k. Tt3p!=#_.?ې62`2mPFDn#W2#ׂxe0DkI:X@S̢@sC 8'20cGJ#»Ax;uRR%> ՉP( .LTIqѱ;jM@&yqp<ΉbOL zSgW|bQ 6s ^M}!s2U ;7d:˒0bip{X)SR\305U< (ϭ*M \y#HQT$Nn%k/nѐΨriP#7PA6iE#( 9ZPZm}p\9[d@ ];Ũkk*K04ă$vDuԴ5;iAF54Hp2wy a(Dp4j`RrQ@{B$CT `,ZP輂rNnQԅE}ێd$\i\5P_\J檅Ƿ+ Jl@Փ 6A!vv.ŸF"c`nDb{A4^#dO(a xsȟ~e@eG'+\e 2WGy%#("\-^4zxE؇D 0& "9SE|[Wc̊ r$lm FN6/3<ʘM2K Β(Isn>0hзYs_:Jx$U`AI{r[%# A<z,ƵTzRo ʠ}O"1/a pAt֕r߻yGZ;Z>ɝzKF|`=Qb*e Jv(a^.$y}'# :jݟ# k4Y]j&%Wt7uԊ6<-}ZX2VRʴng/Z̵Tji PcnZ_.&1 񩺌.LT*] HeIL!Is^EVX f@.#.ٺx i1 d@$zSH^F5U:\9|stPsM g7ևj#]LȧDT >ƕ  p P0N2\o\m&5@4k"8thI˫eUAdm ?7̀3H* VBViO G4/A0+a$Aya,BK9s³nZ*9uLYS$?{#׍/AY*C dXuڲ麐bWVqέ b  Ah:o͸M͐;`-AN}FѰ!k:QqtC1ko]õG.Qƹ5PYSpuKĖX3S)+J>n-ƈg,"^#V&~DjwVםRY ix5V<ؘ uJM}u,z g,=ຂ}Q~O.L#I_i7ddA#sey~W'VhZXy\tRr 5yݣ=3JE(G)Αy5Y^G]Q9s폹hWy9 nvs&'eZ`j»w5ꝧ^*;w&{=7éؽïϜom{ۿ>9ãI }nsCoed[3E=3Ws\ۻ_4"mb|htTnZWªZw1׍~pnnhiS,}5+ YYP}Y![K}S3/kO~) ~mw?|,)<(w{ ~ש)N.\؃GgBw/+3?~C ֥O?O'pϫTk.z{dmMxx͚c^Iݹ_ĸ\WGg2gN׬ -4+P呲"uFc+ۺY(2fNCWZDOchF6ᇿBJ ge;2År:IζInvtjC_<|dmUxn޷o,nݹ3+jJ|_l;M7 ~_Y;lY$q*9jNRx_<+ҩYHV}ZhSZ!rs89=fs2j`S*s @$ 2DͳUʔ5%B;kl'tz:uqNfT_q#5L4ڨE5 f%h}aq\8m9H,wϵbR#q>{_F j5㶶ڪ^@b@ 5{봬Df\_4m(uҲ&'RQi ܖ g)pOaoѮ\7@n JŸln.ӿqwq<˅c.q 8gYWs#knʻS:\sKv:\S1D-jq..ګ` \ dYɉ"0_m gY‡і=4l( ޭ?М1 g>=9/U ࢄIޛ7Aˉ*Y:QkG͘'IbAl|FÌ?VvF3b3؛Wt?m_t/rސ0}4 1j[-euqV+9Wṇjχ EZpP#Qֺ˱>vu]yߞA+=qW⺿xn5=q_Z.+bn\v8 /7gN^RvbPD_y %vE&JQE6,$ƺgq"^2IG=cbg1b|A 9ۈ)UodJ0"~'S: 9pzHӪzدҪTf5iz`N&GO4#JH.(sX&g!:ƶ?M:#P#$o<Vk{kRirȟ>tH[ԋi۳9{:vz`~`{1ێh+dk8̓~ܶ$vj[Jxr]8ZuiX m$LZ `j %Z׷^{IipϢ0H[iH Az{P# |(8r14|5@ "k7,ٱ'J K)ys7ӎ,u؀.xsDS]\m)g5j p&Yzi<|k׎ڴieCP 9WQ(q;܅l}\JZX{,]ӓrc=&E}s;D5=} UJ֎J3wt$[w#ϛcEoS hοCw"SoD;?>9z=Wԑ\ęMT725?{Wo="Y$lf=X5c$ߏT1)JjHƆF/Ū_) hwFVմ{y&ףjj]F* Tj # (+-j9ʊ0ʒ&Ke=b\=wD'P !EY`焕:^***+@HF%ݎ7RZ!g3gR`g0 Kc3'J Ŝ&fiA~{}PGäqMujpbxT^8Kr~F/k@q WRd$J`ӓz+J)oW=&-g}uw+M!6L0I`;.xf+2֫Yy=& XQb#x91cPC(楐$O]a ޮn5_="NEWQ㳻/ G"wSN$RPW25E(ƺO9~g?-&Ccu䕠 #TxHKdɗ':K`I BT9m-#Vum:UB;W`QHWDqPDz^kʖȐKz2ܻPuQ[V:ЖFK ,umXUeS mm1 Ɉv :*Iֈ^PԨ-T6*p0kU-x{bH1\1A-0Z6GZeߔS݌x y40j *4J(F'緈9nEm>JP' jT+u2E.*qn\.X_SH5ebؿhb 6o{K(wM֤Pd8<" (y{3A5?YˋחlGחO7[á")D'B oы69}\ݓX90[(]Otq0\-5nD8R0h..m]vߍu'r\-BDvpS޼fdžkܑzd8Grh@2 o\w\.=%Z)GC=A u"ujOt{0A1sy]]QL'6xq;/cx/3ǯAޥiKD{h(ӣk0@iԠ|&vQ]>7w E3j +(U `G5 T8(3z2@4/*ϝ.*D \TaOỸ5Q-央lP}. l0U8$;T+ʹ8ub5@eL(5(*aS%3y_yj!_1P`c_ܶ6c7e"->t s喥|xPϷ7!S?_^^凋y/rR}^^=e+-WNp7Q({VO0Q6O7~_wg@9UKK,䃛hMw=F)x\ĘN7R颯SGnhޭ &cSS80n21oxS"¬Ѽ[~@wB>Td-n@0Ǽo 1dHĎfDv_6/,)T9iCC15k>'B^kXg:gyvJNQd>wE\S+2xVw.Z?) 6e>QkX+E4#u9S+axsw?M`qaXZU^E6L5`T8$G$J"%@swy&d;U=lrTo*R4q><,WkkV}{Ķ3=zs}ekJY#j9Tj]ltF3u65(LghM.LESe=K)' FfOVD6V`DyZh6Yqw9?.niG .t'px^_87_s5l9t Ycee&SdEZ7ϭj13[Ul&g*PY֤jr%R*V2{"^kZ,Ǩ׬9%ތEM|R"JG1J[ Bsy*O:(^'[O=&r^|AtTn֚Pfv^1였i$L%E!=<7 zI&}/vc*YV8VX!*)0rjlQbռ,+.`&QY[T{IADaΑN$VifP;CeL*[swEDT.TʒQj.>R,2oգ46C".KZ+)$_[s2R %dB,|l42O̵$ǁ/*Ȕ{s- rH_zF^j9/w # =4^%g ۬.VJ˄) V:gkv8JVUZ邘[ 2B~sT-_KiчWnx8a8x 4俵g>c‡+U sdx+{D*TKNi}u_{ QU,"-0Kc8к?x%__nL)xZ^_>߸nu[nt"ua Tky\2r9'fTkKCE]5a_)˟ R)KkOw@Ÿ^VO? ?+ ߾pƞ{W.'oKwqy%nW7L̿m? Q\_=gYFi"qp;,[?5ӣf<<)>giA5 LOap=f&({Cz 5Y.1DtŅ.cEd.QPs=NNx1M)i˽ <^>=sFd It@tH(h#dLJ i~oAEbN;;K B:*r!jH{%Y|l&-!0H:eY=fK"P<+ M1veyVItCz,vsܡovE!T Y [B (*) Y.VF4Åc:^_ZMQ<!t:Op_)>#HjUBGj;p 6GL`"} ;ȅgy""-@%xZ\XۢLU'@(3XxIGUQե9]5UlэĖReh8Ɇ 3ȹmdkFF&;H͝.sIW6y-0z}&(cقV9)ԁ5 jȶ1 C }!ù(i.Qr}楜d]F6lv"|.gn;v`=qq;{)_u ׸#r)r5gFqK`>X&X M#ka[N00y" &'OPdͺ hmjSO㈺m<7'Z?=W?QF:Q\Xb%z꽙aԭs8 ׎F_ﳑYjܣ߃ [G?tHGƛJ`%!j)$k%sPcYqmڂ",%jZS)[y4l_ss5f~=Awѭx'lTk ǜH$GPtVo?>?ĂhJ{ JSP[Պ+bTjdB!k[8b18Lw{ 鏣zk.[>i"l7e\D/6gܮo8fn\<X?7ݜ[Z7y߾2?g˫ݼ_~X{p%"/WՃ[61r5u^om"uK­|\ӿm͕߸??Y[ 8jr}ޓ|pͲ)ߣw˕A~#Ż. :l4 nY65v1nF)x\ĘN7R)O`x[6ޭ &cS:Ҧt{\җ`NsENjM lc3=ԃh[=~ /K%K%$*nuy幭>VqIνNvˍ.-Gi,&1h}á^f',:eA}gҝvVsNy>iz ̱uz㴎j^k9x m35_.}*>{كI) )1C)ѿ7ӗ^)Qrz:(iiyq9 l'aIU9}ׇ2p>[3) ustz cϦt܈7DjܧD(5}ik]y T rq3Jdz0|dn*Jf'F" 'P71#ݟ@'AEb;Z8Skq 6+'&T=x<=֩ݏZaģ1ƃ,E6΍6ӖVʅ5 \( "2]{a9KeW잰T,d>/oC>Ng^<,H+'(ê)I 0"vaMCgt.y&J%(b7I+ 24DeB0ݰ.&'S-'na(,ҎD 3a oS1LpDƈ0  qer|8·Q-@ 0tĥ ߳ﰭ|*k-TAo>ɴ#!̉~9˜N[r~zG'^ڒ5CHyyΑN!W(J=+c $`YQۆqߵ>"/d4EZ)ҥ9<i&O'~%EQC>SH2NR\p3pj_"n6uo+E\" )s&}QWR$B\"=G;K$/h)ID~@?Qg7λi~)ouhC ˳4|*;&OyFxW57-c}XX?ᷢ3}؜vH ={7wo|ǐZ.ʽx%:lO-E sjnkءFs;Pn;$&w%cN haIV`Hbͥfo{~::fQ )P^V܃1S R"dx!ێtZR;kzz&:9!prϸ3LK~uo8NtS&Xd/'3td$ eUE9b|鄡Wn SiaZOĆؤ0%NZI@dFP  d!QKFXP ݭyoYEG'PAq:<8N)\ 2T%{((L󬷁s좳8 ད nW)bWU jGm_~܄3o7/ t^w[ٓ'(k׫rU/@þnXqJhw}pLmlfXrh [/rͳ8ogfҌM'զYm%ym#gz$[~y-i\ԇ76Mnu^*)PVS=3%v0%vˋ/$[<.!#4P͌ }FnkơLޯݎ&}WoZKLۭebKccV"-c0̖43,nf!\+1(lkѸ~{[*q63:G]LL1ҧO:؄n1Sv)zOV4ܿ-KQJ*gq-;2&ƑRjdwBuZä8#+Q+:^3_ gsmֺu~Ȕ;ӝ>p#RE#q%WF@tIgYOR(wz}zo|q`qzE*in1,ܟ֕|w5۰nW{!G̤, )Q@x8050?ZX/Z4?Nb:1jL"%ua\(2iE"/ p!.OgY]:S(WWo]]+{jh3ut6z4l>'>#gs4Jtͬ AgVZ U 1ATӪ8zѕ";.(P'#j+qؗ֍>Mo÷B]PD1 }fBEa`̲k.yY%yK(.$gD_Zr('-@lDkg`Kg`YeY:W@3Mꇊ k\<3mDK\0[b1~nbݭw%vK;Yn%|bl(N`tL*H&ީgv{[=p(`&\ w\"ƹ?z9qNqIB"~p@'X2'{8m<qV۪7fxud3KHtzmd!ypSB܏Z#4r F[IvR d+^'i PUSpfyacO@$==(刷t2\LRe!ZR bTa/XG%v#mڸತESBRbDT"ː(˼ȥrU0*Q~-!_1R U"UVC` 9Rg.@sxs*3Z>JJZJe>hw$SbQ- JUI^Î*Ѕ'z;UqLp 1.< 5N$!W0䙆3p2zK8h$oLmWi1rG"zN`&/mVs O1[vyFj%+m bvvA~;J":y/Ϻ";sl0lo->N0]$`*8/%En&$"S#Rĕw{-  X<Dk1ʖ݊}eec1JGF% " rkntm-Ⱥ X$堲yJfpm2mL`NЧǤG}ZG/Q{z}=i:ǕܟrP"tF w(e Ae#Sd0E}e_F="f_D)E;h(s{VV7B[SK^P"pcC(,dG)؟.Q9cN޳EViĸ?Ms]1J23<;:n1V)0lÓf۲/EKR\"*O_Q}_ߎ"v7.R&YBVvDvޤdu1ՁHbn h*LǒF˽$1+:cB8,RA̚Vðgr0}saKVTe- d<2ܰ< ,y Z=7)9.Z!!?LG]-ݷUMذ~f*mlyu}au2ڜRZL}s3.Mmm@~?S;n,af؋p2 ^Ihh9AX!ٯ %ʿTXǙ( ꊬM_ BK\NxR6kje [r݄5s=ƽ-b^-Fɗ&Y_Dg|N.ߗrz58":"V k$9F-@8SՂK3H!+2ŭ{dZ=汹t+$Er܄"j힧/v͊wN>K90/Hf5vsnAR5Zm(Z?;8+"ʹGa]r2bY}w#KQ IdHiS}:33S&`^ML2EAkU;k%X@ S򐽄ҰkrZk"!IͶb.+ voss7WIjz{ ~?N/_&uS͔ rdV# sCʓ7|#?qmkmKjnE-tz;ǧ#S{e]rgWO_ sq̈́w>u#ks>8L:?>y㻩5$σύpl9`JP)-I+"Rԇڍ+5 }Gv=b*Jˮ[xvkCB^T)yvè0->&튷)OnmH+ъLQOSDZy"qj1$faX;5S1E&D2->pn3Yvz_ki'-LtЗq'.0n?}x{kr#Ol^91HB򚉰JS2V 4Biʄ_8A 0dERn;|=>R yNAv52a)bH(l۷׎ݺ$"A|HAq2LȍQ(RŰ/{ij̆)\wӮƳRDvAu]gMStEQG|̈E؋_5 l&gbL"˾5Kk o醰xt!&z[x❧vkCB^C`e<w4n׻9LK)]k𮟞ڭ y"Z))S| CGk`T'$pI$Ctw]'2GK٩;8sl]:YEsf| (yL a]玽|yM[=3(NA$TQSF@ISIcPTzG?yCr> ‚KU[4|$& ;"ᇺ7jO5c1VƨIחkBSkf)IHlջԳ悢)H %/$.q.uX Z_^܋NG90PnZ$Ќ,t5e2Yj&ŊX^ `+(`fS=RQ)!yTطׇQ7W]de)eǜj.W{}qKRzYjxԥ-XI)oE ճ*Պ&RzRJOJ X%B|RzYjKo[JO󳓒aϩ4ga[R&ti3ճ*S􆥔c/O{ݞHigUEjқR 烐ST*TsT>F)E)h"yF'>zV 2-TI)u/tR'J-LI)sG&0I?)ͩ F]zRJW:{.R:;igU9vK) ?)BJIiN@.q)%OJ se$ 0?)ͩxԥ7.~'ZCRdRw{Pͅ-.e* «Yj9􆥔#/=7\(Q3Z)}(=/,Z3酉6UXobaS\=fu( X>wW:YuR4//vYA\̽ "(eV8z2vEEqw:y?I^>;^i (TrgP QA $=/$%ZHk˗pKڄV!j0dK?ҭ ED f v!`,oedG/{8v \dJ&s`64hhe2;)~mI]~q$O2zewR1qM#ܽ;XMFR!{FC#%5l]|~`PkL7e;.9>aVsYMF.-"{` SPD"-TLI #HEiڒiKCPlΑ\%*?`|~ax`+/| GpEəQ~h:nԓm&xGJ-Qb4j`[fVi`ԍ DZgp`felF+Kt.)nU87{.O\!=|"c*4)T# RiJSpWˈj#4SeR) @zϋZ F^c4j,F\2Kt,δÑ"B"ʘPHi%T3 \OT*K1.'[aIw`(͏s ocwg,6$rk'!o17BO=S7݌@PpG ǟ]'Fr]q)؊'[ *_;SQ0̜8TA7Fea'QyPndWis9+-4N]#AJjpmtoqst ?0jyh6q1U1*r 4E&uF)V|ѡ}{:0>Q{Yj.8ᶫ*sB>AH)P?)ͩQJoQJz%(%[ AǙJu(ф(1"l4tT-$A"Reap&85FD $\f#&f,bGw*~졷fxjvo:TY ? HO~)1ܵ|!/^圔3J%4qX^,|>Rz˳JNdȍ8OV'Nj54.}4ߎm6k{@;W4WB>{\wgw?҇xwP!36>rO6?[w,z.Xr6"'%◗1=NFK&J.fݪZwO}<0R}>ϪWS[ zO!) >e)bzFuW1čY48\)?xxJN9 ykqMY<ԫ3p14=fT(c!dB,R% N$7 m Df%.dq/0$!(W^ܷ7؋hpB_m_ceNA1[ hyGdlXlZf y5La8:/gB#U_/_Lڢx.rm] NMB. 2\"hD:ct@Ri:<&B.m"p9J+cA>EJ5/3Ldi)THS5s_*sKSٻ޶$WY$w{x&ٴ5#:d/~Ք,Qd7Ѷ2@&fTWTuUf7&q>\^f ;h͏ͥßGz뭷޺뺅x6|+ 897aZj)OD3H%IJq!mEcIG!׌+O٨Ldwꋮzr[,gZކ&O읽Ž˽4äˏVlq-CRRw̼!pYCos2KP < Y9&qF.΂RƠZǒĐ$.22)p˞'lB[NoIv1*ɁL)uӽdZVFOm^(h*7Л_Ь#%0Ϛ0sG ~OBe@??o}p6_z7YBo?QAـ8 .z#;_.l{*ܐ/Ӟ"W_T=)q5 \Omn(ٛHK,oN  8RfJȘRPJ%OYUL%K"66HJ-= \ZBC r K׷W$Ktu?QSӑLJx9.~l;GjwNFJU^qB~k<^H8HlE)KgU}I\~HE8 )pgWa'xO B4i1O_ȍd8 ([x?ƭ,L_LΓE¯SIKgVlB9дϳ:rg]xc!{v@y$9P3\]IAM: 'ߤ˫(!o!}>i'Zk Hf@D1S6(C )S`w2)fo(f\Z {jkFS F.) cUBI $ ,9f8N` iMҔpj>L]dz_\:sw SW?3&mN1JyYAL(&hCXM瓳M b.ъRabix4,IF6N9OE&)M$F$O&2eO0򁯮Zвݨv S@ ŀ;~QQ)b""`)N$X J>knHjO(u`5զJX(@:eJ~@lm^2Pa=K]ts10uB2#"bRɅ ؞d$T"Ra{*} >9EN^ꛢ4>ܡߡS;S .@#/ ?G4-/jٚ^Kyk˾}9wvz x>OA`̗`I jmgd煛|妡#`2TvvMmKJf3u1pzkM5rەU53_hg$φ+UG?gG.39SP='/$avJKwv ^ \ oWEhK*@{"W\`z'TC>T]~jozR9l)yla2Y/snPQ6 Zg(@#R*S8@ȋ^eP B pĘSӪ;A\Oy=Z0@uGdĐ闩LR *rsC9oRUORk);juB۹Oy9A/̎Z])JMunkl!w$+(>m\klE .{yXk\c9~X\Wl6_/΀۸S_T}Jrߡ+xٯpI4_tƪzNs+HvUgsGV1 ev/sb༷RyP߹Eʕ& y*ZILĺ1 Vթ2t/i֭zduCC޹V):|j]nU1Q(cN`Y귑[Ux`h}+:}9g9=̦ٿ w; /BΛL>D2@֌e/ˈR"Z7EupYK_ 0g -)Y;LfT m/:(5rc!]kGV;wjT!~~1uTڃZ5H WKtYg?~%"C(:M4>sCKYlϬb$Pw5HC Q\dBљp)"=ơl ]hC``@QtMIxMXKb#%q XV$WNDLk\9X?GSmD ]e!`s^SVTDj-s\r(i$ylN !.JK儤.v1ERȸ'@ D ixJq0LplLRn35 L`V=Pb d'e;~ؖ@nj!g58J`TSe#*ѦE,y:@ z})^ pEx] O*4k-u=mb]V)!@pK}i~UYlO@Q&Fx]5`؀R@(u{F/[.Hk(-3gu;MV"W?z~ >6z5?n;b>0ꋛ/GU`r;LJb\\k80Uo$"LppƃX'=xͿ> dnZKqj%7g|Z~0G bx#wCP(,0X{M{ӚV^2NkZ r3v.H[f<Z+RۖlyvNZXF.g].4&̸bNR&DNӠi1b-4-ɖ2smYsnbԒӴ)$iKOeAQ{bVђ@p؀!TvBWz8o$m9 (OKc \̖F7Fe!`vQ.r&:e/ J *_|Cث5آ0^W{_Ǩ% l&ir}lqHK,yuSXt!VU B U}GDXSDX=АwtʈOL[U BT;X4[`@VsSί8X7)vAT}Gv]Bfݪ_nuh;WтN@IL PwӠg7bJ+[gxҏ4>O G&4ѱ%R&pԄ2g-EM☍\))|[좌Ai%!I\ddS uAP;ra?.'m%q?qxpj}]2ګZ8d"L1 8V?j?磻>l,(!I?Q\vs|v/x 4 i%h@rSM܍Fa.ܚ~s~>~ubkz=t{Yû77Y?{eԕjįf?9hqՂ~'6sA0xG!.5 YJ깱07tns`:Fh/xKTzzlv6 z84|%FMA m 1Z~ oCd_tu'![qIrIEy>Fޯ^ #;q};˶;6m%~JKڄPgܬx{UIb~v$>UN|-8#5$9_Iqm,JR=UAzC Ǘ>u)!KpHp&i$@)]8|s[Q۟ jJq w8P𢧨FQ<=-0a; `&i3Li =;>$)6Tp B!e4x@~be5Zw)E!ΛʈlΦ- [mLRNCy$zi=oPN@hnm2/^IY͛ynm:yR3AF :IB=ҩKy}$v<4G'%T=cȨfspw6&6GhTj@9I"z 1hQ=(8W) 4\:T0-P #61V% xdi$hJYluUR2A"pF`DH! G@DRqOE8m- NBbp)UT?s?5[HǭbS9XY?K |zlF f z n&}52Mp4;O>V2q^&Ϻrص/,rw \/ӫKj@?{gG{_+QoO{Od՝&8;m맺ܗyY!Gɏ* \V-޵5m#뿢MvyړS]vSǩ)^@[kH5x]AR/!Qh%oX VȞb &}EHAR`yD3E0޸7&;-h5E^PG^T1YLNkxܔO.Q>w_R˂ 4+geI}6$^<ܧ\ * I^RV7#kQ[%D~̍AZjN =_P&ED  +ؚJ1BpȆc񋯵l׹>TVv\P@O؞*Ym2gKr$D$F>e{AӃ׌H>,_'cy/00|-Iu|- SEp,cPD58PGJBz:ڣ,TO| ?TW =#h FT<dmr*La2xltUq)  E jBW8`&P#Q4QTt=r']o C"wRfqܕw藲(*~蝓!.vl]DbEZNݜ}p{S/T4CN cqH{ N+ѧk)J:Z)#D 7_ӵ呜6H.]dQ4Gy.-xÕ;ZgT.~UZQOM6k6+]{H8Wc|RSӾNa\KQƄsvFƇ2q&R!yFF2Q̖0)Ta[_Vyz]FQ↥=0"=5_Ix$ƫuFx3TDErպ1R}{+U]kdUbZǕ;ɩ] s$:{ׁ p̈ʝ\95죙OVg&هG]'֛uZ"b%SޗRHU8ôfU+A9ɭ!X%Q$!%I#b\䨽Bn:G+ND2}+p%!a<؅V.Z¥=-0E6k.`|\孕[+V.oruFoO?R{Dʀ"n(w!LObN3dW}N֌'hf`!bOCdL?CХn35>ӦwQ'{>C6(>*|KUժ(@&b]璾j7YE:yx%esciP΀J{ ~ v`.n4GAe!f 13z @3"ZD_(Zw[پ<)GAoU,0B& ?^ϯVO1~wD{m y%9f멻٥UyBH[<ևr|-/oOAju uW(Nca5cmk"q*;yv}bM1& ̔j)tw,`"XhBazx.vZb=A_1<T+_K8}*#{ yHDF@4/&8#l6;T;]{|%Af]zݗ:x R2TaH)PH+"J #G[DtzP19"*t^9%( U_FUB!ċl+UۮK}v%0UUx)jˉh_* G@ȝ7E^h" i1C^⏣>]gBxw,K)Nwon>yyǫ(oʓ(;/lvAo \wG")р*%oYd,3@Z-ĕJrvX"!8`ʤ`zH*l~i$C(x瞏W LJueݩ٠Znv 6!nRJȀpN5= )}uҳRJM@jkR6[deӏ?=ct5`ubS¤Ez"X&4bIWqQˤx< f3h_ "r涍)wځޤ9~|6{ x1 /_OfspI}#<˛7&=ޘ?F3cs_^Fw/G<vdϟѝ}yz'A 9g3 _ݏ?#oˤ-{12{uo䫬l2'>h)1l&2|}2 2&J2, whtmZ(nuaCVGxƃ i})V# ?̟~]AK*MʄWSsw5֙?EbOo~S#YH5 oM íFZXj= 1D+:T˅"m$O,E&O>ySvPGvUpnד(<)yXF+_{Sϲ ,L0j׳#h&Lrt竩%%Np,@S(}d2 Gi? /oǓO7ā7N!d&+ aOBop)C&wShgzA{2|b/LN/k$/~tlv73 /o=?? ^.Se1'h5L(|$&3x綇OYH(unvU .fBqIKP„԰>j-[%RRwK9E;MQ6S T~9KKkukfOxMiHeSך^00hEJT/5iodnۗ%3f=)Q?j 8jeCj]_ϻI^ ^)=RpXm xN2/px9Z%<bkǐmt隇˫iaas;iƔ ^)ʔe"sViZ.q'$emMJyޟըwe|Fў)ْZ+ݩ䓶#z<j)dkTv{a hz RJ$}O}nhށNѐRX3VGz- ݢ,y=[gFW<娄.sTʼn>R1Av7mdy)`ڰS.4hL:HG?[RGIZ=inI"pcu۝%+QF߃wv{DOqCVyg,)htM{/Vեg/n0~I-[}(Fmzt.n%+G95ss;+JMt4eWq@Cb1 S9I;9ٗ8AAD`iI"&fǕ c܃G2[Ѩn(k7Fƍ:[S(4'Q鷌9߾[5y[.{<$5ag^\EY 1>{3(͌"RvJJE9 xi=e]1:2d!Zʩ_93V:L7.L~v'WvXrnAD~~1SU VwrBD]i؀)oXILk7:*,k˚k*To{WQK,ƫG4rr7Gnz2kuMt6qs}iOߠe!'r03m/1l$ǫI.&%Gq9CxMϏ&@^b^9ʨ8 [6rsi9en8]f`#s*&Y)^Jo#]TmwT tv𗤳 y琔}E`n&w.FڦX;{CnSdɥ a6BێX5pJ8ϗ1@A\iN=礷r0TNx}݇•Z.\;P2|F6p\=Ԁ.?ZUZ5U3+6K }y{vs8{gJ׻@h3OJr(Q"{'؝/A%|IÕ4>&>3q{K唹$YFTWsMR(c,9hNb>RsrNon!P("MRTNLhQ·bN{qZ!9 O8DNPUrrimx,*)}j,+Б԰䩱^j95Q֥ PEfFqCj65TF$be Y%cq3f}jfpcɁ[_gF7hq(1P3t?iR*b[q> )} :ƀ:IfٽOKSP_o~;~鳸V‡`Q8$>ߒ21[L3;NF* 0ZrA%]QZݔgx=.[ ~RH<ЯLdS+*]n}#)n[)AJl%#<\f_ELPIr#@ɚӆi A)yi?v<@@ :Nig!13΁28}xDUBOv=$")!y3E"'!Ǚ$X&AKe<@ʠjCΩ`!C (} D!FSr9rȼ 2d8Ӛd*g@=qXlEĠ$˝#utF uDZ|+0źl)QskLFCt!fFˁCdݖ b3 !d % ³u.(*iFq2㊣MQC^uU)E?԰!unL9R]zQ$5G_T7:&J)P;` 0f2mN $ Z Kqh$hI(6=H[ TNYP ] ,`3y!Ӳ D(d؃W\j-Ҳ K. d#߼0"&D- +шN-p񍽫O]LǤE-yZRUTߗZO,V IH/ Ls #V @94"TL rj2&x93aZy*DJx#Ê8\g[z *E k v?buS!JոRyX'֒7TG}J{@tV&y/КQtthbnq03¨hIO7sc6 (lGAy4n˥6҃nϥ~ $wk,\ؼ'; mzCOF .o楚P􂓔IyNĮAVxRڐZFZJTX,I5-Uf $WW n]Ԇ~;`k߹ͻIod=㡨i]/aVaH~vAӯcP\HఌN@ϤĨqJD;>m>^V#k"U *̎YYxnM%-`]}דTJ=A1\cS4˼߃wa{g1E"r![~/~Y3Ѵ+7_0e&?]?^^ݖ5ϓ_#>م=o)yGf o8*S jǠʱє{ڶeD.J$&WEPOC7|lo Tʂ3-00Hң'O2Z #iݗ tZ3W?N yׅN?,yå<`rcFK"ҀdF$HETYIg2S+,E56)غ6pZ&;ȏprf7m}&a *!8;_o^5!Aۧ!q6.0ˆ!Gf/547GɱᗻLL22fK k A7RY]R-h^sj uu[qF7""( eH~6x?*~_ݣg;?rg˥Ϣ$)ϖkAj.E6)˘gB\2ìی؎;GHf +M=^!#m_xOYqUsV_Zrio盽ޢn~s.>=wHz}ln{s.>ۯH)9^&nD/xx`ig[t>տt;fNuǃ7w=3$qʬQ<7 yF[πk[ m' hػ^4tR_aKl/>Y_=k'zr wgqmNCY"q+dzFd눨ۻm]v-τ&wW69-k[:iy5Ϸ-d|[S6x^.Q;>뭗p[PfmV%&:}3OfН$.dۣk7Gݦ`L};z_7n[)n=$.d٣8nF4|G6cDgw30j硵kFjG2#e OKF&tz~cϜtjlŝl:QGҬz>3iOݮ)fY[4Rti{n[(ongY[~KO.~r|여֍*y*[e w<bXEoLEX#}G*&rgGV Ov*P+rYok rgn2\}1gd'0th l!EYx*e2RپI|(ft0>d$[hIj8ycܜWf}Kpʰ'C,xDZyto8l.T@t y bڧR]ZC,ΰoИN̮7&%s];1hWI0jd@ĉpDs{`R{nW-cР sSoBefMHX}ضmBEo@jU`Q6e?NUcÔR@J9Nk RX/;QDU7T):-=d=E{ff=>v^yOFUo&L0y3g ;=dK>>d;#LhѫO^~^̗!CjV8D6>B&ֈCqqOH<ʉ{QHv/7Z-uJm> ;W^{O@ED&@Ԟz~TA/(P䵁 n2de;Ԩ ת7ij ;3 Եt=r*蘉0 v̳gr.ck"ݜ.ۿ9ڌ7XBYRj;m?n1|*%2x|I0 (L453+,`t%B(Nff4mHڨQHyߟNxhegGiP+'G ضIyQ X[us;P{SmQ煿-^Ie:Ok6A<jlU:Dߜ>f\hP)Ԃ jr bI dI71Twʢ\mV"K3M k6ZcjJMbՐAjk A/?+z$6}"!8ުyɅ6aܡY!`Qɚ hcYMvp'*k ugsX^;2Ua-$*̦;bAOiϚx^#a:eEREW>ƆBDVH(d_ٳTR6fO{i*Z^FߢE%dSI%cd -~?@# {;3揋t{((dR GľȐF2.ŪI_~8"=1ؤ:IfA_%U}*`s*=ТU &+3&^"$ۗU9cvC{bkqn 3՘ ]=٣~p*8lg b4D%ik˪x)UPDe={i/O4#a @";b驸T;MG̸Y#:{)e7w_w*EI5ɂ'`kvnK[h>%Ya*<㼖&,jq0GUW 1Ln!DSԳG/Me#aY#GSgmۛ:bp!TR#c*>W=hbe8ͩXb="j 2y{$)ʞX X79v$^{"\^i@嘑,Ҷd3 ZS6mNUӒWtf~s O8D-D^[dtVceTt)[~8-AoK'F𪇌KU!5L+20Cn0A(d+8vMz_r*5CAݹ}xH5Z+Χ:*@Mк=|XJ+gYjb<\|61V2H:4bpdbo+ٺl.~Cm*>o uּח6'B RVt|aH9֙{ a#䵡_NvQ^jU3Uiەc=f@tIY!VS94$^!T6^O [7eY.Ʈ ):J"U,()Vn6~4QInZy&՘YuD x$ } oFJBG%Z]FgŠj4Ĝӥ.˾# YtNؔ`ڧ/VPOYӯDo:kvyTEqc ƛHыuZVA idw=^TP;RSP;|!v<>J>ŷWY2KV++z(h$ClwlhSqrLSdo,WᴴlDDqwҬlIM6lcǻ,.ddrxZ( W٬=p5u9oxCZs x3b !H^0zd)'²ڥbq/pׇkZ[ݻD`W9l: 3ZcP }Q{ #}Dj:Г1x`SeSuZ^{v֪NE$ I}JkqBv73 t-O$: ߦKK{&\Z 賀Bj4lx*e$7e..zҶF~tQy@h2PFʑlW~[]-ۉgw@OL)ziVTP<*c+$dố_>؃pӶ:ݝݘdݿ3#]sbͽ4ڝmk[sO1ʂNs[LJt$e#wr9xAY4D8N}b4PIechI)K XźsHvNk2\^{y'aKdpuΧud(7UHN <@w" c(U'{1T)@hE[6I({HpwK]3qh]O/XtؑJq=da "&r/W#!\秋="b%6!PEXnG΂0*7`}_Y903zPU Z /*VB+!=y!OdTLa] S\4 !VmPr-kt,VEֻl+FoPV]l aAgI:M$rq~|_^|Ƙ cbV^5IEK;E]sƖ?nnH"L`T h%h5K;D*GpN|:(JA"H=[·0*{ >I޵5d鿢ew&E/IJ蘞m<;*jd{$D!U2;g}%py_2@yEvqp7zV=.Qww)>JIh7_KU 1ߗ'1芣ؘ-\EA|FB3T<Ⱦ/sCDCsOV+ڂϸ-o1U>w^;6O7d\kat R4L=WEؘĞQ| <'ѱ_Ǡ&;,=p4pLN3z4^=)O,{Ajw̅8`s>e*[@P|+E Eew *v Lx%pnF.0(`ˢJrwR3 gt|ap p!TI"&fzw?9$k]ʵ^!_1Y! Ria.oG ƔtVoh$]Ad?ô;sk b ƶE Xwu?O\ ;'a~ \gE!4_5(ĝ x,=^fyOc/<9 0A"O D r8&ZPN6͒ HgS8 V!F `":pb[oL6*).h!Ɓ[kG B)XRDAYN{v7DMd([.x=(]@>jZHEN ?|Be~-T+ݘӻt߲~0^ɣ{GR" d?QXݷ\{wtMeU,ӥ=#U]{[1_If^.gFt~<8錇ӑ;WppLʘ첌BaGzRꞀ)m@9 ~8R |WR?"+鍡<r=k_!ȯ~.GpC7. t;YYċpWc)@)\q=h>n@4XC8(OyG@ :v B㨗ZL~L߇Iؑ9u>{agaL!Al>w41thiKHEWrD.FsPJVe?s Uz2C4{&!&$PXi"%Xp7-Aj&KihmU {an1[L&+/͈7k!&;O8^~oKì]GyW"n|ϊ9Kb1\5N#8qU*UZMnZդ &OerEs7PL?j;p-g6" )<8Vr?Yt$tr|II0as?F*I^iP#7HS1̇a^Чs /f6!FDQ"0G'ڥ2BF#·ȹfrB"UM5/3W?c8(J!@u)i"̩ԚE+vh- vv՜:p:ޯDv:rifzSwL;Qї4k0kE 3B^Aha5#@FtE]8ɟy1gAڟ:g?@| spf-!蜥cOG`_t? s3D^*0~]| p2^.)e`"i+j2j.jk |}µāR伷KQ٥\ ْCݷlgiX x\CgCEe@qށNNL%U-ݫ{dnuL}'eJ-9;LdWݤ$oǓKzGk bI@p P}`meX$$W[3[{]Q~Eۥ6Y>\á;An#x\ȇ }yTU<aHru_^ӟp4>M٦ CQz./MG0"ceǽ >]"yCuВL&zU St+L ¬Lx?8٢4z;o6!bc}-$3612oAJy[-45 wE;:^ Vg\d84ќ5Ԍ^9_rjΛco0{~ tf- l|ho[m[sq0*r4/huO`s, p CFJ<c!Ӎ}sd,RLq/z<ǷقH܂p?z1V]3xyFz$箴ĕ$Zm빮yHX|>jfpEu셇ZBIQRҹ{Zo~VsC(ōiݴ1`ݚҠ4ccYjZ6$;%y܉.\(fk2ӫ~Yz^皿[k&BU{nw,=N ۹OBvQRdC. ND4Aj0(IrƯW)½A 6aN>_0*}*dpRK$sx,&ǰKaC7Y Hj 8t*tLiP"|JͧZ!)4Q] f*2Qgr)&%5h]5%LL' b~wucM70BBd)(FA' #cmBb+FIjwx(=?F+ܡ9=Sxk= &:#@=@ m1C-8fK4̘\o\q' S7:/ 0r`-Ҁ -&2ݥVeV+/IChu}\l"#7Zdq&/HwY ^@؎P״OIQCYFOg<I(Y#Y%ҹ 4sVB&d5<^(SnC&ܤ4^.#'X9a #A3#O+Nui%APKwW7_~S-upх+{/r'qE-yG8Jt-0֘pa7$2?: <†q#PN#CXWY`|])Vl"TF{ I(T &!" $S鄤ǁ#ODMhu6JHԃ @ȶ- 12W \j)Xr'U&$Sa}H?BֈoK +GRJSB0b8g/2%)Ϝ~sB@/ !`Äz@${Q + ;!6b9ZS̵gS#19笂H mJa#?NDK,I) _|s8`Y[C6ϝmhBR󭪆%0DٜoV5 ZJ qX<{R"R/qC5D$ F&LL )7&)\QS`B6IAo McT)Cߏ,+c wΙ:[vj5Sqb ٤ONKӧBrwmXȼ_ ĻXl 1CrmdK΢%gdY3cmgx<Ui2K4RKS*.//ﳺ.-2zJ]&qH_vh3X-^S,YHi'>ߍgK u2\O,w(F 4{Sc.Z\祫㳫K9.CᙦH\!^%PU #0 >"m .fCY@XF{-b)DXB Z EPMl]ҾOJ~HrNri yB,=.|(aYA ]uk`7}ҳ_|m ,#W+!iڲ3+E@sQd[r ߽}*v8(WC}\̷eh<UKCdS.ȫVUe _;No/*CXA,Yo~Ъ.|sgUQ\?q ޛ?Y/NW7U狽 O-˭H6k(=SzP" 5S(zB*$1.^EEofIİzf2(<&}755;{7ޗ b!Q.aF2 )A`mC-`z-'CV<31>d1X?|\z+`sZgnZY@h[z7N?L'kiiޣLIoQLHbti8y7ceUC4>׷b=NS4N"n*sstڢHbgIuB&A]>S}pw^5G*Nt8PYq;$C/2;t}4i`iJ.s&q\ En44Tpb C-r, '͇3YI,b?t֛w",q*y'-e?0PaJo_/Thin\ TL[:>_"4\Vjҩ cW ;=Hm,= w~jܸ_it-]p^~ؾ' xyuk8)^|4_7Q, ~R(pm'=KS CEF]_H \t~.#~F%G abg`fYg{,"eXZ3vw!VآrXiիrszmjk Pk쒞믓OsSvOǹ2 U O>̂\jz++nzqշSwqQlTfHߟGNӟ@~Nux_WVl'ߺE 4suu4H6*cDJ7-sfNf hE{ʂWul7O2 ,%ET9FBh#U J%@CX\l)D.)kPBͰuvn~:92'V;K:JscܻI°2t!ݒ."aҾa#IϏ݇ &ibEa"/ZO~%soŌJL('6+ xxIz"S!W-%3Vx*k_>#L] I(Q@)p D֖rCsCiA8A( ѣlP.WBR[5_( sT" ”+9əYAR`ݿ$x6˚B2r/@=c1RWp, 2@ty͝eI7]j[(o1gl_CDvv^'Ö!EGy)MgƬyNJG4I zxSNl*`wya/yaӝ~Iš1*Lʣf OJ0(|\zeV쁴w'jihΒ6vj }ZEA}LCLX &䇟^˅6nͽ;dUb41! rLj)1 &:I /z dB?*a(Q\J?υ1ēۛZ$EDXT9S @UZ}܋QYea+Va u*_HQ@ .u>r.40ԂJYr(@@DLAƄ'$Qe?f[n0:WY:=3?q|/*潫=!c%JC=VZ ,]Ɠ #4DJ<Bpr5;>/mx($~Di0 ?%cv8( XS.QN!ZFjƏZs㹂T#%K)sLP*xYBXڇKdI`}Bii0XYdfǜgVi Q6vjپpBSVB$T,Q?PV3\I-/ $ Yv>IF_O|B&KV{#d7ՏΣ'n̽ !lֿ!7<6>vHX*D-^x;~g|3Z n%E5A* *>Z{0j4\[S/Zx(ԠXt+@c}NeU ZYJ+pk5; ۼrV@sY?1d>,aY3~]Bg!_ l9텱Qn\gL^byw{klp^<(lG=W#Q t) @ 25 he=`l5vk?T!@i1hM)W6Pz~eC9vOw]15/X=}[;jt /ۇ?]!oTx31_4vyWq;0ݿǟޜX2愼R[@ A۽ZdPB`Tt8+9H->XtXx{sc[>;AWއIsURޅu/Ÿ/y$p3Ap@;FGY;D!YQK9(®$ϵnTa#H$rh񱓇Qv5coȚR p4;۩|UiS$0R$v%пTIKkt]D´6$i')FH]$pZRL`:gl!v%66U@ K:gadIcV6Y~Nf$𠥐]T36&K==1ϑ [9RAoʑztԨflq'طҹN鬟HوQڴdNLXԶ yz(#C,ﰹstoŔ] :F>dXJKmKF9 "˿J}BJxn!;f~Q#fvumfĨ77Hp&p1~Ne}\oE?wd.кАW1:!=xsź!EJѺb:hbVYg->!кАWK:4L)WKEև ;u~>\ևÁ vi cdX`pT doB X! 9h2vm/F?1oų{Hb;C_$m+TtI:y=],)H-Q.gDJBJ3ref(Ӹ9Lڹ9y`#t^X{ViA,E~!@ "gM˦E(WG~#h7+kn#GEٙ82q8b^3/(\ִ&壽1}$-)RBd DA97׫>PG]XсG}Wt9Xu.?*?sw`M'Ak7㌒tm'Ou%z$t->gjjLy1$/rP,!0"9/ Rcu6T;IcetE3/e$WgDzǦf5AXjtm6_/,JͻdKOaFe3hL8bo64dpʀzڎ,?7ɍ0K!=@P49 dߠd/-2 Jⱕb# "f|$3𽃛QepoOz61; ї؀WF5䒍N`6V`m:>bԒ.<0ݎiQR]xR $V-"6mB%=|L |$ WJ:ͅ@ ĽHKZ%6$A㩑u\JuiNn 2+1+kN a+yL$(R Є XZW'ϲ2fAu<ʇ#_{mލ_IhQ8>HWFҎ`-וW n9I{4Ph\B=eJDz,iZ`WX.}t-n'cNB=xZ<B?͚yڄsٻ-??7+v'ٿl7b{[W'0u}CezYR>wutaw^y$+h@Ir8sHhr)hKF&ECvٔڪx&KmBkx%~IV'_sآJň[B q~Rf歷nk|nss!R#\xCV%w>GgbZZ{RҺJٟ-*x63e;uGΓ!XЭ7 V*V8>'n>؈e>?R-}ifjj'l)ξ"~%p57]_4Ϯi405 y4¿xπSvb^pFLßnӾ,Tן$ de&@anmXa=Ns )>d-ÃjЎg<]%Ius`Tip8F^\XG]r`tҁKyh&J"'ȹƷZ9+V[AJ%&P'M I/M* e.-])9/mDAIsb*O!T=dp**I| zgm)5Q*+%{ < K:2^ˆ>Cړř4P$M' P0ȳ^sfY`hn ℳFmlģ`^kf+6#\*aq(S++/mPd l._۠G\z7`1o~v"T uHaw}ोwGV~Ǵpa ̀<ei1%hrȻ6%5+ 5?VvjOSA(^&=Yku ݋? q ^,~MRZ+>/Fh<u)"{ > ?zHE]Xߛ򥤓8gJ=O/CyJR&9~ߔ<)J%O'm%FkL+eJJ9h45 zc)\Aр*^\" xbXð@1QbX$%oC}Ti9_S+{Js0#T}:ɊP' z=eڪ"a8nQP i|G ݽCeHqe(,ܻDte酘jIn̂ ,niS?DMBҏݡiaoy_?U7TOn8.5@hG'8ҽ*\X.V*޵Ai]?M"T \?V!ONCf;kXn]^]P]4, w,>~3C5d׃ߕӨ+e+p;ݕANvgƾ*Su3}gPJu](Jt#McYF6vqH07RջQ&rq4M}Zb>}Wf=0}7Ҳ!<1^qG [MX=-mUO n-'cqjMQ;dA@ԒEνdZC`wy_*}V/]} U񌧩 Y YZ=Pd͑Z^,qm8'mz_uC5#lE0Ƽ}f/-@!O(y~+tT̾)yR<)J%OJn$&HL2*H.Lxh4Z svxj3[WUo>|@'Xk`X N ]4J![\lwfݻ; GU GQ׉;BCh\U \HǕYg֌*w0 Ƴy3xIe|M|3󗡹Zm1y 6#hϜ҂FЪpkv~Ke)|`1L/{{\4n&"GH|SM1ók<xJFZZx!]N?rg֓9r Ln82{'1:kyKjg=5wRof/HUW׷fjF_+W a"0mdIsIɒ\%ɇW6s Σ3 {ٸ0gSRvrJR#M >hӁ=J ?{aFYPO&PDLLr3S" RHRo,bmO=sm +!.ÜXIyB#s9q/Kxp˟C'Krv[5uc|ܤ~N +SW9SL-陫qaJE*҇E&K0kB "Y'eR6hD3@_nՅAJo~\ȴW77Zd0m )X̔V)7.7NI{=sHۦnQ~I< zS򀌑 ֘>٫O*ɤ`hMdrL&Qy j]ɏ Ν>~vƈо/l8~B#YBO'4?6Ny 0D")r(^$E$je0WU]]U]]5fMBeqL< \ 4kIEio㳫eW3&rsZ,m|5m?/Nq6J&[g|KǦ 劉=B o9OqP G/Bꎹ&=Tk"W;h,\#0dsX2gn?ڪZo XIH '^dX4)Nz_Ji K>DzS)[r֫I΃MYmŔ";O=rN]$/փb!<VI*ݰo7ť0WmqQ笍#iOi9\3\nڶDjɴ$40#H}mq~H,{! pHQ:%:rֻf\&. a3$ ټȍ>Mf(&Ȩ(ay%'cLѽbK_Z`$ ܈ lֹHѤl^.HNT#aX|G$r iphޜt?s! bSt XjTb~֥\יrrdF mWs 3v`Fӫ~BStvFF7myтhN3"QDXԷ vhGSLc01R"G0H!EZJS`x kH 839AbdEj/$W>TSKEZ֭[*u =z|}˥%|[B;ᶯ-Nۙ]\9+xJ$Q䂀56BU6A8ALP9-G{s8* ; ?l ryHgϽ 7g4L)q@,W5b'-@BC֎"euWd):(dIcv ly4)|X\}(MjI""u܇d2Q<&?i-Sֶյ*2$ӜRCP\9]rhJnMqbny]|%*c,UH- `> BVuoMfas_}t\A̹q_jE>P11"'<u Q2&'0Rv)eKm_jΜ[g/MP$M(Uf^"Wt3_DX?9F8Ll7[ Ɠx auFKb{6XR IlaG+[!A \5kVKĩt~$m>;kI04XMח˱tEn5[[ͼO0GL~ )ϦoZyw{)79?rSw$߀ڶ 5hvmGVe%4f}YJ׬ o!zRPlY7?}ւJHJ fc:&d; uQd~sC$to=c$Kh - wvr<nΙhIVge,8֫dIٚS[yWw;p~,F|r1on!rn󷟝+QE>|ά5TUi'MFe}!e0{2${!vvoT3GA Gev+hАMiћߎ;X<z8~<#R+MfVz.\\p[rK%,yny ¹gܱ{r +ϖ$;F2ee%[SEK9ٴ[얚ڭ ELUJjvCnMiPGtc?- ݚ'$nmHw.O)MiaD2޳NcE:'=v;d9$6O&^tVDt {*^4rWC;fN)]L&v/qP>4_YsK Tm~dEޮ̋ lFƝd|69:/v1Z{zW/j|OHX~ 50%*lFZβc AaIdڠ$|ښ(/Ҧ0AJkYܽGUp`'^,6܁ wO&į ٠9a-dҰ F2Hu}A2v&C˦МҖܢBW/v"5eBt<(0iA 4e/1J-vި bo6eK\>`LB q6L"(aFu܊׼%18aq?\:6FEP.>Mep5Dj9&􃓈Uaa1E[VZg[e_'P zxYPB Are"FzF |_Ϣ-bě D"Z !\G\X,X|HI>yy h:zD p'4q_ $dʂ$KfLxfkMs;7a૥C, %P① P$dYsky|orߟ}WޠK*x!鄈 NP v u+BO (Ei9>򵓁K׋X%4T'ڂOroL?|$0wNy_2D(3t\ia{z4d'}v{L'R5?>JIk7"@4"7J ѶK ^aGB.]]8d(},"Q(W5Wѥ}y so,Jv`ǃ5^* FFG \Ǡr)Y^Dag?v=Ӛ)ly/ A1W|Z09<(ʬUրQNx [-Jd|FϪa> ?8`7v1\0AZ=XTIVq97>#_Wf *){XTI"FXƘݹf7ҩ^l.ΊE0zKXm !m!CnQoB//{Ʊ O'uTZmUҦR8/k&іH=eDQP3HzyhAե܃Do_RG}l!R ַ?@ٳ@hFZdcD:}$t6V #|+ȉ/G w0h7$lSus`}{c}(~$91mQĪ#+>|^R |mЛSb]es˯}h? s)E-<3+Lǀ}]mC.dž:`N}/P~u~Oa _[ZضvAYӾRmBO ~jp+"g0ʊ8dM__B6{6V!ƱaW*z*a,Cooq8/g%A?f+vKFc6M?I^g 5Dv#c#/&S,D.IJcƫeK -@ أg?I>Dž N(hI*4-)4{ˆZ(vA=;\R#vzZY hYľ?ÛN#:85eI+1}G?٢ @ͻ{LK4{[:IW/} ן~8W.8D$1J~aE"@-JP ˟6E^y+k'$7^[ T`#pO\ǃAN" VKxa z_: !7&ϟ1"RFFX*HEloqƐ%Wa`?wO}WdSe,|A HbAAboakƵ|~ņv\s(J!s(ѡJT-}Ebn}ZKKu=VTԑɗ4+ye &/D')AH*|F '1ʹ =:ͽwV($qoo%Wvq|02zq7E(0*X8 $*bЍ5"4KwKhmj$|{sj-2ZonrY!,cǞ&*聤Zಛ-i-y ,>*!)%@E b܂n#e222Z|vOgQ^$D;iGƂ$Fe1е " |حױ]}(Ae"?+Hkན 1xYρ|A*&INe)ҀT>j8CPcBN^H.McA~=Qסl$B$x+6(ORy=\q1KfS8ʀ4z3$uDMH"W^.c0GT#ѻRz TRPivk's?qPO69:G%U¡jϒt_PXr۩c=(cHУK1X".T;Hמn%F(mA-g= Wbܴ] ovjؚ= |#՜4Gq7ϕx@LiU/Ma f2VS)\9j)^V|FT|^-g z˹5F*-cG^t 9+Zס%G߿~+T.'"$eD=yE嚹{=n55 %|>8wW@+3iltk10֣,9=N[OzzްPKɏrsd>rbpWlG Y YM6ŠXiB*|n–pf(>hJs̻#5EYjT)ml fWV Ex5 zVcSiRk{s߰|RE!HP?$/BH$Z%l6ׄ{[-jL 5ԫWZVL5El%GׄsV97%kV*ڷVCQ%!FkZ*k.]-}Eb;[ZZ\Vro'qZ:8eYR΂pxLF|PR89RfWÆp>GX}>_>A@g*igQM$f]qIIkŅUF"HtrZ2'Kq<&wųrl#(P߼}{vI>R=8bĚɂH$XG?ÛAe)19:Ź PYR-} dBp>b֥Cx`w>37b~ipR`Lcmn($5{]/u|:7=VeAUrkԵ]zL8|p*n@@] @=:rZ˝ѷ[8B ~L8` _Lw6l=R] V &J0c?-,G= @ޝw&34 1$Pm ,c\VY2w(7.Ю kmRd8ϻ3??G?~Iu%y6-|w)SRA( HO )*$ZHA: .Uۉ7h(yO:@򪼮?ZsOѭtݯǸ|9xuVp{K ~ܘi@M+[[ s H!zR0[@Y*8K7@ @٢C[-Р+1PSdZ&)Q/5vz\|,""bCQg) -k [" $ԖC&eς`#uwZy2K ΀bDʈ2*  @ VF̳@@Օ PtWwl?)8 RG?bGWwuI<mRfoڋ.LqyǛ.dz)xoJn~9;g6a=SJU}\wqTMX2U4v.?s.@{ATMS=GS}@ʏhjjiQ%F vk9\ t/b̌&$ӓ/MUT{i@O&7Z֤.Lfu/U/uSf֡wOp)/̉-NfsͦgM5zsc>aNNw_O s31v`m waA3%՚1nc5man]m(hŘ_&BLy˳o9z<|V(_ףîĎglS/nh$AW6)ʉ]t1tf -2k|Q-A'1J?\0\B/QX uui]"7<ۂ‘zWX^6_(+lf>~Ve wnqG,ƇvKCN:G?̍.9[s:?ap{p=w#83z2 K}q^ pJ36xØķ'ujV6V{wϮxHҪZkͻz(kSw 7f,p/^*HlFV|r?/BCM@nx|j,>ấ +2†+0n]"4VߧSm)'RݸTTËnDTEOZ{sⷣ X !`ݶA^{?tydЇjʃ?vuM1<σQ4n=2tk!OyT.ײF_a:C9yӿV2e+^h6{ӷqZF(lK4.SHjRPI\RF %8gWej6c\☜;c=ŹGAMyDPIc5=`Dh,ʾG:ޛH: rRj[PPmGm+P쌯1>P'QQтVw 8v./Y9;3C&g GF~O=¹Ӭ0։4̬Z tAeJў Z")K,PONp# c@tM Qw;ORr.~ Fj5@m3&e }b7]6 瑆4Bk޾'mraუNp w_c#J)/p`{պ-57q(ý 8J~~A4pZc G@T8G*0ܺ+-۠}q|{!^A&iOG_J9 хS99ӑC\ ey"=*nj|F.Q"Xg m8iEBr*GK&Xj5 E*͹ƫXunHNxL dsQu=S:لbm;w:#H#6RK5G;?v}ph>ܫ 14179ms (00:03:54.339) Jan 24 00:03:54 crc kubenswrapper[4676]: Trace[660571960]: [14.179800913s] [14.179800913s] END Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.339210 4676 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.339544 4676 trace.go:236] Trace[791521244]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Jan-2026 00:03:39.635) (total time: 14703ms): Jan 24 00:03:54 crc kubenswrapper[4676]: Trace[791521244]: ---"Objects listed" error: 14703ms (00:03:54.339) Jan 24 00:03:54 crc kubenswrapper[4676]: Trace[791521244]: [14.703645735s] [14.703645735s] END Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.339557 4676 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.340464 4676 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.348726 4676 trace.go:236] Trace[613796798]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (24-Jan-2026 00:03:39.353) (total time: 14995ms): Jan 24 00:03:54 crc kubenswrapper[4676]: Trace[613796798]: ---"Objects listed" error: 14995ms (00:03:54.348) Jan 24 00:03:54 crc kubenswrapper[4676]: Trace[613796798]: [14.995151288s] [14.995151288s] END Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.348753 4676 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 24 00:03:54 crc kubenswrapper[4676]: E0124 00:03:54.349778 4676 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.421322 4676 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58964->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.421323 4676 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58968->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.421367 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58964->192.168.126.11:17697: read: connection reset by peer" Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.421408 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:58968->192.168.126.11:17697: read: connection reset by peer" Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.586133 4676 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.879450 4676 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-23 23:58:53 +0000 UTC, rotation deadline is 2026-11-30 21:36:51.557210319 +0000 UTC Jan 24 00:03:54 crc kubenswrapper[4676]: I0124 00:03:54.879496 4676 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7461h32m56.677716848s for next certificate rotation Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.161120 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.161656 4676 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.161722 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.167994 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.196931 4676 apiserver.go:52] "Watching apiserver" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.200258 4676 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.200607 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.200957 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.201196 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.201348 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.201424 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.201924 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.202064 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.202130 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.202334 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.202485 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.203772 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.203913 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.204064 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.205338 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.205416 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.205427 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.205478 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.205938 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.209241 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 09:13:46.822899023 +0000 UTC Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.209429 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.226194 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.239144 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.253663 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.267496 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.280610 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.299911 4676 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.321438 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.335863 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346677 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346719 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346745 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346765 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346784 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346798 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346818 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346837 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346857 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346875 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346891 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346911 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346933 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346955 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346973 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.346987 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347002 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347020 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347039 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347055 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347072 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347089 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347107 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347122 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347137 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347152 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347169 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347194 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347211 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347227 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347247 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347269 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347291 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347316 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347345 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347389 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347414 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347436 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347458 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347479 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347510 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347538 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347560 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347577 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347591 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347608 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347628 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347643 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347659 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347675 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347689 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347706 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347722 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347738 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347753 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347767 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347782 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347796 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347822 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347854 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347870 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347886 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347901 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347917 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347930 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347947 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347962 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347978 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.347997 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348012 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348026 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348040 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348054 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348069 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348085 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348101 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348115 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348131 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348147 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348161 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348176 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348191 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348206 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348224 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348246 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348270 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348294 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348291 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348315 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348332 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348347 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348363 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348395 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348411 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348426 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348443 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348457 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348471 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348487 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348503 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348521 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348537 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348552 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348567 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348584 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348606 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348628 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348646 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348663 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348679 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348695 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348711 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348728 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348746 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348781 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348796 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348812 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348853 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348870 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348886 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348903 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348920 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348936 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348951 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348969 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.348992 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349015 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349033 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349049 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349065 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349080 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349101 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349116 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349132 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349149 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349164 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349180 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349197 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349218 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349234 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349250 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349267 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349283 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349298 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349315 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349338 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349360 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349389 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349481 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349501 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349518 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349534 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349550 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349567 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349584 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349601 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349619 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349635 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349653 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349674 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349709 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349733 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349756 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349780 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349804 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349825 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349848 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349874 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349897 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349919 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349942 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349963 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.349987 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350008 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350031 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350054 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350079 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350105 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350128 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350156 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350178 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350194 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350210 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350226 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350244 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350262 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350281 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350299 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350318 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350337 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350354 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350387 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350404 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350422 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350438 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350455 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350492 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350512 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350547 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350564 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350582 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350602 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350617 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350638 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350661 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350680 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350698 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350716 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350734 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350750 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.350781 4676 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.351250 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.351626 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.352312 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.352798 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.353055 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.353550 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.353799 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.353937 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.354134 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.354217 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.354501 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.354905 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355046 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355141 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355167 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355235 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355234 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355393 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355453 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355475 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355805 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355902 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355955 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.355963 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356101 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356145 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356250 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356309 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356423 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356573 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356661 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356691 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356858 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.356886 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.357010 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.357080 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.357102 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.357268 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.357370 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.357534 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.357696 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.357736 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.357754 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.360255 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.360534 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.360560 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.360657 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.360805 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.360803 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.360939 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.361069 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.361310 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.361429 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.361617 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.361796 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.361929 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362072 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362203 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362216 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362387 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362467 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362541 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362619 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362647 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362736 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362759 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362774 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.362953 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.363077 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.363168 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.363606 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.363974 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.364496 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.364714 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.364926 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.365120 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.365301 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.365570 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.366117 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.366195 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:03:55.866176269 +0000 UTC m=+19.896147270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.366591 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.366728 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.366806 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367097 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367139 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367233 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367420 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367483 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367573 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367575 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367748 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367921 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.367966 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.368186 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.368515 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.368708 4676 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.369088 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.369359 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.369606 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.370118 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.370297 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.370646 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.370773 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.371160 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.371541 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.371879 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.371941 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.371901 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.372643 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.372657 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.372744 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.372765 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.372795 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.372889 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.373002 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.373196 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.373614 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.377225 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.377327 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.377496 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.377657 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.377793 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.377964 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.378119 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.378270 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.379023 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.379266 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.379314 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.379485 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.379566 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.379798 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.379983 4676 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.380032 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.380059 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:55.880042962 +0000 UTC m=+19.910013963 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.380125 4676 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.380160 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:55.880154335 +0000 UTC m=+19.910125336 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.380778 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.380945 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.383428 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.385219 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.385520 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.385997 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.386348 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.386649 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.388699 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.389081 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.389292 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.389428 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.389721 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.389900 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.389908 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.390154 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.390169 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.390449 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.390481 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.390510 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.390529 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.390678 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.390717 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.390791 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.390807 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.390818 4676 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.390871 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:55.890852867 +0000 UTC m=+19.920823868 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.390993 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.391003 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.391346 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.391441 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.391656 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.392538 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.394682 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.395084 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.395212 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.395227 4676 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.395298 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:55.895280826 +0000 UTC m=+19.925251827 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.395352 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.396178 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.396412 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.396465 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.396553 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.396626 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.397040 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.397108 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.397472 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.397530 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.398520 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.398628 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.398932 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.399241 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.399406 4676 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc" exitCode=255 Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.399428 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc"} Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.400110 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.400589 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.404776 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.405652 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.405760 4676 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.405993 4676 scope.go:117] "RemoveContainer" containerID="5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.406017 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.408002 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.408011 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.408457 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.409691 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.409824 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.409980 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.410073 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.409660 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.410156 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.410207 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.410209 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.410266 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.410423 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.410489 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.411424 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.413175 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.413504 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.413765 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.414278 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.415134 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.417794 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.417970 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.427627 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.434995 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.441133 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.444211 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.448218 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453154 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453182 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453240 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453252 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453262 4676 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453270 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453279 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453290 4676 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453298 4676 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453306 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453315 4676 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453323 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453331 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453340 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453349 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453357 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453366 4676 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453389 4676 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453398 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453406 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453415 4676 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453423 4676 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453432 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453440 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453450 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453458 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453470 4676 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453482 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453493 4676 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453505 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453516 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453525 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453535 4676 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453546 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453558 4676 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453570 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453580 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453589 4676 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453597 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453606 4676 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453614 4676 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453622 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453630 4676 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453638 4676 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453645 4676 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453653 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453661 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453668 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453676 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453673 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453810 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453684 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453848 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453860 4676 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453871 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453880 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453889 4676 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453901 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453914 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453928 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453940 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453951 4676 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453960 4676 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.454155 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.453968 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.454756 4676 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.454773 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.454784 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.454795 4676 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.454807 4676 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.454818 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.455351 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.455388 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.455400 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.455411 4676 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.455422 4676 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.455433 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456160 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456176 4676 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456189 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456200 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456212 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456223 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456234 4676 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456245 4676 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456255 4676 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456267 4676 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456278 4676 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456290 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456302 4676 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456365 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456395 4676 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456408 4676 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456420 4676 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456432 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456444 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456459 4676 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456472 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456485 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456498 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456509 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456522 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456533 4676 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456545 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.456557 4676 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457014 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457027 4676 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457093 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457106 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457141 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457156 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457171 4676 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457182 4676 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457193 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457229 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457241 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457254 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457266 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457280 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457318 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457331 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457342 4676 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457352 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457396 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457412 4676 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457422 4676 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457435 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457481 4676 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457497 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457509 4676 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457519 4676 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457531 4676 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457542 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457556 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457568 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457581 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457591 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457602 4676 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457613 4676 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457624 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457635 4676 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457645 4676 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457656 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457668 4676 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457679 4676 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457690 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457702 4676 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457714 4676 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457728 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457740 4676 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457752 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457764 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457775 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457786 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457797 4676 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457808 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457820 4676 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457831 4676 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457842 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457854 4676 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457865 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457876 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457885 4676 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457893 4676 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457901 4676 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457909 4676 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457917 4676 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457925 4676 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457933 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457941 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457950 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457958 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457966 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457974 4676 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457983 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457992 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.457999 4676 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458007 4676 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458015 4676 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458023 4676 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458031 4676 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458039 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458047 4676 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458055 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458063 4676 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458071 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458079 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458087 4676 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458095 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458103 4676 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458111 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458119 4676 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458126 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458134 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458142 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458150 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458158 4676 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.458168 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.466158 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.476267 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.490965 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.526303 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.535463 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.542395 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 24 00:03:55 crc kubenswrapper[4676]: W0124 00:03:55.562211 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-9284ce85b5ad52e51c0ca7e0622038442ef0d0993981770cbcd7b91cc77b0dbd WatchSource:0}: Error finding container 9284ce85b5ad52e51c0ca7e0622038442ef0d0993981770cbcd7b91cc77b0dbd: Status 404 returned error can't find the container with id 9284ce85b5ad52e51c0ca7e0622038442ef0d0993981770cbcd7b91cc77b0dbd Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.569140 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5dg9q"] Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.569417 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5dg9q" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.575355 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.575582 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.578283 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.602126 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.655769 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.664632 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/efe79b06-a59d-4d3c-9161-839d4e60fb52-hosts-file\") pod \"node-resolver-5dg9q\" (UID: \"efe79b06-a59d-4d3c-9161-839d4e60fb52\") " pod="openshift-dns/node-resolver-5dg9q" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.664664 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cht5r\" (UniqueName: \"kubernetes.io/projected/efe79b06-a59d-4d3c-9161-839d4e60fb52-kube-api-access-cht5r\") pod \"node-resolver-5dg9q\" (UID: \"efe79b06-a59d-4d3c-9161-839d4e60fb52\") " pod="openshift-dns/node-resolver-5dg9q" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.668357 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.689692 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.710993 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.724135 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.736025 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.765671 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/efe79b06-a59d-4d3c-9161-839d4e60fb52-hosts-file\") pod \"node-resolver-5dg9q\" (UID: \"efe79b06-a59d-4d3c-9161-839d4e60fb52\") " pod="openshift-dns/node-resolver-5dg9q" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.765733 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cht5r\" (UniqueName: \"kubernetes.io/projected/efe79b06-a59d-4d3c-9161-839d4e60fb52-kube-api-access-cht5r\") pod \"node-resolver-5dg9q\" (UID: \"efe79b06-a59d-4d3c-9161-839d4e60fb52\") " pod="openshift-dns/node-resolver-5dg9q" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.765890 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/efe79b06-a59d-4d3c-9161-839d4e60fb52-hosts-file\") pod \"node-resolver-5dg9q\" (UID: \"efe79b06-a59d-4d3c-9161-839d4e60fb52\") " pod="openshift-dns/node-resolver-5dg9q" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.774750 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.789206 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cht5r\" (UniqueName: \"kubernetes.io/projected/efe79b06-a59d-4d3c-9161-839d4e60fb52-kube-api-access-cht5r\") pod \"node-resolver-5dg9q\" (UID: \"efe79b06-a59d-4d3c-9161-839d4e60fb52\") " pod="openshift-dns/node-resolver-5dg9q" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.866366 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.866550 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:03:56.866532825 +0000 UTC m=+20.896503826 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.889218 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5dg9q" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.966848 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.966885 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.966904 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.966922 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967018 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967031 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967040 4676 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967075 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:56.967063729 +0000 UTC m=+20.997034730 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967116 4676 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967144 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:56.96713893 +0000 UTC m=+20.997109931 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967168 4676 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967186 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:56.967180861 +0000 UTC m=+20.997151852 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967221 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967229 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967237 4676 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:55 crc kubenswrapper[4676]: E0124 00:03:55.967255 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:56.967249393 +0000 UTC m=+20.997220394 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.974552 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 24 00:03:55 crc kubenswrapper[4676]: I0124 00:03:55.991484 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.000432 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.027766 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.038079 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.057069 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.072814 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.078897 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.088960 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.100313 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.120976 4676 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121108 4676 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121130 4676 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121166 4676 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121193 4676 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121192 4676 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121216 4676 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121238 4676 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121246 4676 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121256 4676 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121294 4676 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121259 4676 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.121320 4676 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.121591 4676 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": read tcp 38.102.83.27:48472->38.102.83.27:6443: use of closed network connection" event="&Event{ObjectMeta:{kube-apiserver-crc.188d81e0930901c4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-24 00:03:50.165971396 +0000 UTC m=+14.195942397,LastTimestamp:2026-01-24 00:03:50.165971396 +0000 UTC m=+14.195942397,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.210144 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 00:53:49.465445343 +0000 UTC Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.255269 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.255396 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.258630 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.259122 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.259850 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.260985 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.261563 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.262035 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.262947 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.263476 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.264431 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.264890 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.265747 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.266340 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.267167 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.267696 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.268257 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.269081 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.269588 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.270309 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.270838 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.271340 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.272124 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.272730 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.273113 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.274037 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.274527 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.275454 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.276010 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.278069 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.278114 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.278636 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.279481 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.279911 4676 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.280006 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.282756 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.283227 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.283725 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.285146 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.286086 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.286564 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.287511 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.288109 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.288892 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.289441 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.290363 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.291293 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.292055 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.293284 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.293405 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.293947 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.294776 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.295218 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.295706 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.296143 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.296651 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.297209 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.297668 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.309280 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.321446 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.339172 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.342469 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-x57xf"] Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.342716 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.344000 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ld569"] Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.344571 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.347232 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-ppmcr"] Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.347477 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.347734 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.347903 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.351111 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.351299 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-7mzrz"] Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.351713 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.352205 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.352395 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.352535 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.359945 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.360046 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.360220 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.360306 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.360336 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.360451 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.360455 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.371942 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.381163 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.381799 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.381921 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.382199 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.382681 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.382726 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.412947 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.418299 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a"} Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.418536 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.420490 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.420857 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a2ca09c6467fd8af4e726359910c882217012224cf8db9e93fe056f2f8f04336"} Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.424616 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90"} Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.424643 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a"} Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.424652 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c11a4d1f7d9392b187f98967c34b669b246804eb46906bd61cfc2be965a4c421"} Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.426201 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5dg9q" event={"ID":"efe79b06-a59d-4d3c-9161-839d4e60fb52","Type":"ContainerStarted","Data":"53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b"} Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.426242 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5dg9q" event={"ID":"efe79b06-a59d-4d3c-9161-839d4e60fb52","Type":"ContainerStarted","Data":"1116159e810236268e73eac6cd4b12bd9b570d8e39f251e02cb4ad8117e01696"} Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.428034 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51"} Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.428056 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"9284ce85b5ad52e51c0ca7e0622038442ef0d0993981770cbcd7b91cc77b0dbd"} Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.435555 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.450740 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.468586 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471296 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-cni-binary-copy\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471346 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-socket-dir-parent\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471364 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/bd647b0d-6d3d-432d-81ac-6484a2948211-rootfs\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471396 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-kubelet\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471411 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-var-lib-cni-multus\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471425 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-var-lib-kubelet\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471440 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9vrg\" (UniqueName: \"kubernetes.io/projected/bd647b0d-6d3d-432d-81ac-6484a2948211-kube-api-access-w9vrg\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471456 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-system-cni-dir\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471470 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-ovn-kubernetes\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471485 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-etc-kubernetes\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471500 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovn-node-metrics-cert\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471571 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-system-cni-dir\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471669 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471727 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-ovn\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471745 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-script-lib\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471777 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bd647b0d-6d3d-432d-81ac-6484a2948211-mcd-auth-proxy-config\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471793 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-openvswitch\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471810 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-run-netns\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471910 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-daemon-config\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471945 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-netns\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471963 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-log-socket\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471976 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-bin\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.471989 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4frqp\" (UniqueName: \"kubernetes.io/projected/24f0dc26-0857-430f-aebd-073fcfcc1c0a-kube-api-access-4frqp\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472019 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-var-lib-openvswitch\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472033 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-run-k8s-cni-cncf-io\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472049 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-cni-binary-copy\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472071 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472105 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-var-lib-cni-bin\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472121 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-hostroot\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472136 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-run-multus-certs\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472151 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-slash\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472165 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-config\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472209 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bd647b0d-6d3d-432d-81ac-6484a2948211-proxy-tls\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472225 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhp4b\" (UniqueName: \"kubernetes.io/projected/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-kube-api-access-rhp4b\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472240 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-os-release\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472283 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67bbh\" (UniqueName: \"kubernetes.io/projected/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-kube-api-access-67bbh\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472313 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-systemd\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472348 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-os-release\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472363 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-node-log\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472405 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-env-overrides\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472420 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-cni-dir\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472434 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-cnibin\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472479 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-conf-dir\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472499 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472513 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-etc-openvswitch\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472548 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-cnibin\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472561 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-systemd-units\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.472575 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-netd\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.479944 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.488908 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.499227 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.511036 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.520939 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.535339 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.546013 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591307 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bd647b0d-6d3d-432d-81ac-6484a2948211-mcd-auth-proxy-config\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591346 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-openvswitch\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591369 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-run-netns\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591418 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-daemon-config\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591434 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-netns\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591452 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-log-socket\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591470 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-bin\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591485 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4frqp\" (UniqueName: \"kubernetes.io/projected/24f0dc26-0857-430f-aebd-073fcfcc1c0a-kube-api-access-4frqp\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591506 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-var-lib-openvswitch\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591525 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-run-k8s-cni-cncf-io\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591541 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-cni-binary-copy\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591558 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591593 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-var-lib-cni-bin\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591608 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-hostroot\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591624 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-run-multus-certs\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591712 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-slash\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591739 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-config\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591755 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bd647b0d-6d3d-432d-81ac-6484a2948211-proxy-tls\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591770 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhp4b\" (UniqueName: \"kubernetes.io/projected/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-kube-api-access-rhp4b\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591784 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-os-release\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591799 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67bbh\" (UniqueName: \"kubernetes.io/projected/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-kube-api-access-67bbh\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591814 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-systemd\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591829 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-os-release\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591853 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-node-log\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591866 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-env-overrides\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591882 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-cni-dir\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591896 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-cnibin\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591909 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-conf-dir\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591923 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591937 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-etc-openvswitch\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591953 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-cnibin\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591967 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-systemd-units\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.591984 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-netd\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592019 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-cni-binary-copy\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592034 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-socket-dir-parent\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592048 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/bd647b0d-6d3d-432d-81ac-6484a2948211-rootfs\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592062 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-kubelet\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592129 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-var-lib-cni-multus\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592149 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-var-lib-kubelet\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592288 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9vrg\" (UniqueName: \"kubernetes.io/projected/bd647b0d-6d3d-432d-81ac-6484a2948211-kube-api-access-w9vrg\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592329 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-system-cni-dir\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592345 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-ovn-kubernetes\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592361 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-etc-kubernetes\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592391 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovn-node-metrics-cert\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592417 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-system-cni-dir\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592432 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592457 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-ovn\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592472 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-script-lib\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592661 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.592883 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-node-log\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593333 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-script-lib\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593421 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-kubelet\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593444 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-env-overrides\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593473 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-cni-dir\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593469 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bd647b0d-6d3d-432d-81ac-6484a2948211-mcd-auth-proxy-config\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593497 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-openvswitch\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593512 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-cnibin\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593527 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-run-netns\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593538 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-conf-dir\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593730 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-var-lib-cni-multus\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593794 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-var-lib-kubelet\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593972 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-daemon-config\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.593991 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594016 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-hostroot\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594027 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-etc-openvswitch\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594025 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-var-lib-cni-bin\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594054 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-netns\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594085 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-systemd-units\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594089 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-cnibin\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594073 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-system-cni-dir\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594111 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-netd\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594120 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-ovn-kubernetes\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594135 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-log-socket\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594161 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-bin\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594151 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-etc-kubernetes\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594458 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-var-lib-openvswitch\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594506 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-run-k8s-cni-cncf-io\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594790 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-config\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594821 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-cni-binary-copy\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594865 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-multus-socket-dir-parent\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594892 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/bd647b0d-6d3d-432d-81ac-6484a2948211-rootfs\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594042 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-host-run-multus-certs\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.595018 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-cni-binary-copy\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.595149 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.594063 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-slash\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.595402 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-os-release\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.595421 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-system-cni-dir\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.595436 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-systemd\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.595467 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-ovn\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.595507 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-os-release\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.597630 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.598293 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovn-node-metrics-cert\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.598846 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bd647b0d-6d3d-432d-81ac-6484a2948211-proxy-tls\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.608319 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.612441 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67bbh\" (UniqueName: \"kubernetes.io/projected/b88e9d2e-35da-45a8-ac7e-22afd660ff9f-kube-api-access-67bbh\") pod \"multus-x57xf\" (UID: \"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\") " pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.613339 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9vrg\" (UniqueName: \"kubernetes.io/projected/bd647b0d-6d3d-432d-81ac-6484a2948211-kube-api-access-w9vrg\") pod \"machine-config-daemon-7mzrz\" (UID: \"bd647b0d-6d3d-432d-81ac-6484a2948211\") " pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.614633 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4frqp\" (UniqueName: \"kubernetes.io/projected/24f0dc26-0857-430f-aebd-073fcfcc1c0a-kube-api-access-4frqp\") pod \"ovnkube-node-ld569\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.615868 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhp4b\" (UniqueName: \"kubernetes.io/projected/79ad333b-cf18-4ba3-b9d4-2f89c7c44354-kube-api-access-rhp4b\") pod \"multus-additional-cni-plugins-ppmcr\" (UID: \"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\") " pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.624178 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.635057 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.651295 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.671440 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-x57xf" Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.682413 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb88e9d2e_35da_45a8_ac7e_22afd660ff9f.slice/crio-8f476b3316aeee36dfa281df4104b2636e890e125d49340c30276eddbcd20c67 WatchSource:0}: Error finding container 8f476b3316aeee36dfa281df4104b2636e890e125d49340c30276eddbcd20c67: Status 404 returned error can't find the container with id 8f476b3316aeee36dfa281df4104b2636e890e125d49340c30276eddbcd20c67 Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.682800 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.703272 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.714468 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.742130 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79ad333b_cf18_4ba3_b9d4_2f89c7c44354.slice/crio-5cd2eacc7418e20bee23791d1168847e95ffa559aa7a9f4b13fabb531e7b4d1a WatchSource:0}: Error finding container 5cd2eacc7418e20bee23791d1168847e95ffa559aa7a9f4b13fabb531e7b4d1a: Status 404 returned error can't find the container with id 5cd2eacc7418e20bee23791d1168847e95ffa559aa7a9f4b13fabb531e7b4d1a Jan 24 00:03:56 crc kubenswrapper[4676]: W0124 00:03:56.765084 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd647b0d_6d3d_432d_81ac_6484a2948211.slice/crio-6792889cfb51332982504d10488c5aa6324a4ef991ad23646663e4d7cb7468bf WatchSource:0}: Error finding container 6792889cfb51332982504d10488c5aa6324a4ef991ad23646663e4d7cb7468bf: Status 404 returned error can't find the container with id 6792889cfb51332982504d10488c5aa6324a4ef991ad23646663e4d7cb7468bf Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.894856 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.895046 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:03:58.895031668 +0000 UTC m=+22.925002669 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.995813 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.995859 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.995877 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:56 crc kubenswrapper[4676]: I0124 00:03:56.995905 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996000 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996014 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996023 4676 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996062 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:58.996050281 +0000 UTC m=+23.026021282 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996108 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996118 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996124 4676 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996142 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:58.996136503 +0000 UTC m=+23.026107504 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996168 4676 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996184 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:58.996179585 +0000 UTC m=+23.026150586 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996218 4676 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:03:56 crc kubenswrapper[4676]: E0124 00:03:56.996235 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:03:58.996230766 +0000 UTC m=+23.026201767 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.018678 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.038464 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.210265 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 14:12:14.519097545 +0000 UTC Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.254667 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.254702 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.254787 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.254859 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.263348 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.273546 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.366039 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.431581 4676 generic.go:334] "Generic (PLEG): container finished" podID="79ad333b-cf18-4ba3-b9d4-2f89c7c44354" containerID="0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe" exitCode=0 Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.431645 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" event={"ID":"79ad333b-cf18-4ba3-b9d4-2f89c7c44354","Type":"ContainerDied","Data":"0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.431668 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" event={"ID":"79ad333b-cf18-4ba3-b9d4-2f89c7c44354","Type":"ContainerStarted","Data":"5cd2eacc7418e20bee23791d1168847e95ffa559aa7a9f4b13fabb531e7b4d1a"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.433863 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x57xf" event={"ID":"b88e9d2e-35da-45a8-ac7e-22afd660ff9f","Type":"ContainerStarted","Data":"db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.433886 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x57xf" event={"ID":"b88e9d2e-35da-45a8-ac7e-22afd660ff9f","Type":"ContainerStarted","Data":"8f476b3316aeee36dfa281df4104b2636e890e125d49340c30276eddbcd20c67"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.435890 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.435941 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.435951 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"6792889cfb51332982504d10488c5aa6324a4ef991ad23646663e4d7cb7468bf"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.437162 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551" exitCode=0 Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.437195 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.437260 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"a4ecac14869f9e61b6cf328f1a06fe9b463abccd1352f43143e709699b48fbd7"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.438395 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.456200 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.475294 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.488474 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.499649 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.504651 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.523779 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.543535 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.550262 4676 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.551856 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.551887 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.551896 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.551994 4676 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.558301 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.560105 4676 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.560235 4676 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.561007 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.561030 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.561039 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.561054 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.561064 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:57Z","lastTransitionTime":"2026-01-24T00:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.568953 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.572687 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.573959 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.575489 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.578478 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.578533 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.578546 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.578562 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.578571 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:57Z","lastTransitionTime":"2026-01-24T00:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.586196 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.587580 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.589159 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.592144 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.592170 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.592180 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.592195 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.592428 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:57Z","lastTransitionTime":"2026-01-24T00:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.598274 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.602967 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.605336 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.605361 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.605370 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.605406 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.605415 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:57Z","lastTransitionTime":"2026-01-24T00:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.608936 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.623106 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.625685 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.627679 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.627773 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.627850 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.627914 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.627972 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:57Z","lastTransitionTime":"2026-01-24T00:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.638094 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.639509 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.639658 4676 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.640886 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.640972 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.641035 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.641093 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.641144 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:57Z","lastTransitionTime":"2026-01-24T00:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.650222 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.658253 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.662794 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.667308 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.673054 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.676149 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.682462 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.691205 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.702216 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.716236 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.728900 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.743472 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.743528 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.743539 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.743554 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.743565 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:57Z","lastTransitionTime":"2026-01-24T00:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.744399 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.757188 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.767867 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.776476 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.795821 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.846242 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.846273 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.846285 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.846301 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.846312 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:57Z","lastTransitionTime":"2026-01-24T00:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.885581 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-4bcxm"] Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.886058 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:57 crc kubenswrapper[4676]: W0124 00:03:57.888886 4676 reflector.go:561] object-"openshift-image-registry"/"image-registry-certificates": failed to list *v1.ConfigMap: configmaps "image-registry-certificates" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-image-registry": no relationship found between node 'crc' and this object Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.888936 4676 reflector.go:158] "Unhandled Error" err="object-\"openshift-image-registry\"/\"image-registry-certificates\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-registry-certificates\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-image-registry\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 24 00:03:57 crc kubenswrapper[4676]: W0124 00:03:57.889705 4676 reflector.go:561] object-"openshift-image-registry"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-image-registry": no relationship found between node 'crc' and this object Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.889749 4676 reflector.go:158] "Unhandled Error" err="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-image-registry\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 24 00:03:57 crc kubenswrapper[4676]: W0124 00:03:57.889704 4676 reflector.go:561] object-"openshift-image-registry"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-image-registry": no relationship found between node 'crc' and this object Jan 24 00:03:57 crc kubenswrapper[4676]: E0124 00:03:57.890202 4676 reflector.go:158] "Unhandled Error" err="object-\"openshift-image-registry\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-image-registry\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.900400 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.914341 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.930803 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.948473 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.948510 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.948522 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.948538 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.948550 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:57Z","lastTransitionTime":"2026-01-24T00:03:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.953716 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.972022 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.984019 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:57 crc kubenswrapper[4676]: I0124 00:03:57.997169 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:57Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.006280 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzmhq\" (UniqueName: \"kubernetes.io/projected/bc086f6b-af67-49e4-97c8-f8b70f19e49a-kube-api-access-gzmhq\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.006324 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bc086f6b-af67-49e4-97c8-f8b70f19e49a-serviceca\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.006343 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bc086f6b-af67-49e4-97c8-f8b70f19e49a-host\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.012883 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.026035 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.038406 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.050542 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.050571 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.050579 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.050593 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.050601 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.051368 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.068023 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.086567 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.107498 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bc086f6b-af67-49e4-97c8-f8b70f19e49a-serviceca\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.107546 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bc086f6b-af67-49e4-97c8-f8b70f19e49a-host\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.107632 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzmhq\" (UniqueName: \"kubernetes.io/projected/bc086f6b-af67-49e4-97c8-f8b70f19e49a-kube-api-access-gzmhq\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.107810 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/bc086f6b-af67-49e4-97c8-f8b70f19e49a-host\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.127642 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.140454 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.152439 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.152460 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.152468 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.152483 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.152492 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.156791 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.208961 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.210596 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 08:30:46.94354132 +0000 UTC Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.213790 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.254662 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.254692 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.254701 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.254717 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.254729 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.254824 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:58 crc kubenswrapper[4676]: E0124 00:03:58.254922 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.258116 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.295344 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.330294 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.357132 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.357158 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.357168 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.357182 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.357192 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.367567 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.407190 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.444347 4676 generic.go:334] "Generic (PLEG): container finished" podID="79ad333b-cf18-4ba3-b9d4-2f89c7c44354" containerID="dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9" exitCode=0 Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.444409 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" event={"ID":"79ad333b-cf18-4ba3-b9d4-2f89c7c44354","Type":"ContainerDied","Data":"dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.450841 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.450875 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.450886 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.450898 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.450909 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.450922 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.453536 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.461201 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.461234 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.461244 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.461260 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.461270 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.488677 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.530333 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.562721 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.562747 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.562755 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.562768 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.562777 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.568070 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.612904 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.650914 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.665438 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.665493 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.665507 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.665531 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.665544 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.690369 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.726004 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.767668 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.767723 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.767737 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.767757 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.767770 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.774894 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.806598 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.839893 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.870553 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.870601 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.870609 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.870628 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.870637 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.877219 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.910651 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.918670 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:03:58 crc kubenswrapper[4676]: E0124 00:03:58.918808 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:02.918782965 +0000 UTC m=+26.948753966 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.955331 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.973397 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.973448 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.973463 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.973483 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.973500 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:58Z","lastTransitionTime":"2026-01-24T00:03:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:58 crc kubenswrapper[4676]: I0124 00:03:58.989813 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.020011 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.020251 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.020242 4676 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.020360 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.020501 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:03.020474659 +0000 UTC m=+27.050445660 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.020349 4676 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.020588 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.020697 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:03.020636364 +0000 UTC m=+27.050607365 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.020745 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.020765 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.020785 4676 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.020824 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:03.020816629 +0000 UTC m=+27.050787630 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.020940 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.021018 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.021090 4676 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.021198 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:03.02118409 +0000 UTC m=+27.051155091 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.029341 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.066439 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.076439 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.076511 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.076526 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.076552 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.076572 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:59Z","lastTransitionTime":"2026-01-24T00:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.108492 4676 configmap.go:193] Couldn't get configMap openshift-image-registry/image-registry-certificates: failed to sync configmap cache: timed out waiting for the condition Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.108613 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/bc086f6b-af67-49e4-97c8-f8b70f19e49a-serviceca podName:bc086f6b-af67-49e4-97c8-f8b70f19e49a nodeName:}" failed. No retries permitted until 2026-01-24 00:03:59.608587267 +0000 UTC m=+23.638558268 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serviceca" (UniqueName: "kubernetes.io/configmap/bc086f6b-af67-49e4-97c8-f8b70f19e49a-serviceca") pod "node-ca-4bcxm" (UID: "bc086f6b-af67-49e4-97c8-f8b70f19e49a") : failed to sync configmap cache: timed out waiting for the condition Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.151172 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.172295 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.178989 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.179083 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.179098 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.179123 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.179143 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:59Z","lastTransitionTime":"2026-01-24T00:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.196424 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.210815 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 07:11:08.689485911 +0000 UTC Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.235637 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.239431 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.249834 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzmhq\" (UniqueName: \"kubernetes.io/projected/bc086f6b-af67-49e4-97c8-f8b70f19e49a-kube-api-access-gzmhq\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.255145 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.255248 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.255311 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:03:59 crc kubenswrapper[4676]: E0124 00:03:59.255357 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.281601 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.281643 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.281653 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.281667 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.281676 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:59Z","lastTransitionTime":"2026-01-24T00:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.289051 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.327472 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.339886 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.384329 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.384601 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.384705 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.384802 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.384898 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:59Z","lastTransitionTime":"2026-01-24T00:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.385788 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.440412 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.457858 4676 generic.go:334] "Generic (PLEG): container finished" podID="79ad333b-cf18-4ba3-b9d4-2f89c7c44354" containerID="6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5" exitCode=0 Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.457902 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" event={"ID":"79ad333b-cf18-4ba3-b9d4-2f89c7c44354","Type":"ContainerDied","Data":"6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5"} Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.481189 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.486928 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.487121 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.487249 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.487413 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.487498 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:59Z","lastTransitionTime":"2026-01-24T00:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.512299 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.558516 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.587852 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.590050 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.590079 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.590091 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.590109 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.590121 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:59Z","lastTransitionTime":"2026-01-24T00:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.637192 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.641042 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.644910 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.649835 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bc086f6b-af67-49e4-97c8-f8b70f19e49a-serviceca\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.650694 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/bc086f6b-af67-49e4-97c8-f8b70f19e49a-serviceca\") pod \"node-ca-4bcxm\" (UID: \"bc086f6b-af67-49e4-97c8-f8b70f19e49a\") " pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.667979 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.692231 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.692268 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.692277 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.692293 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.692302 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:59Z","lastTransitionTime":"2026-01-24T00:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.699480 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-4bcxm" Jan 24 00:03:59 crc kubenswrapper[4676]: W0124 00:03:59.719543 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc086f6b_af67_49e4_97c8_f8b70f19e49a.slice/crio-422ee2aec88f6d9695479975fa9bf7b2c3d3eab9e3d87f14e41463e52aeea3b8 WatchSource:0}: Error finding container 422ee2aec88f6d9695479975fa9bf7b2c3d3eab9e3d87f14e41463e52aeea3b8: Status 404 returned error can't find the container with id 422ee2aec88f6d9695479975fa9bf7b2c3d3eab9e3d87f14e41463e52aeea3b8 Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.734284 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.752420 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.788513 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.795345 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.795410 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.795422 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.795444 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.795455 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:59Z","lastTransitionTime":"2026-01-24T00:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.833323 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.868972 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.901950 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.901976 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.901983 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.901996 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.902004 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:03:59Z","lastTransitionTime":"2026-01-24T00:03:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.912777 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.946223 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:03:59 crc kubenswrapper[4676]: I0124 00:03:59.987903 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:03:59Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.003921 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.004088 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.004145 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.004204 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.004273 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.027046 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.067333 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.106508 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.106574 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.106595 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.106620 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.106637 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.110080 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.146713 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.191615 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.208817 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.208854 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.208864 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.208880 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.208891 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.210967 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 11:32:35.258624193 +0000 UTC Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.225990 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.255512 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:00 crc kubenswrapper[4676]: E0124 00:04:00.255638 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.275699 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.309687 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.310940 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.310994 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.311004 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.311019 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.311031 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.349931 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.390654 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.414050 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.414090 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.414100 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.414116 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.414127 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.430560 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.464775 4676 generic.go:334] "Generic (PLEG): container finished" podID="79ad333b-cf18-4ba3-b9d4-2f89c7c44354" containerID="1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44" exitCode=0 Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.464810 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" event={"ID":"79ad333b-cf18-4ba3-b9d4-2f89c7c44354","Type":"ContainerDied","Data":"1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.466862 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4bcxm" event={"ID":"bc086f6b-af67-49e4-97c8-f8b70f19e49a","Type":"ContainerStarted","Data":"f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.466907 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-4bcxm" event={"ID":"bc086f6b-af67-49e4-97c8-f8b70f19e49a","Type":"ContainerStarted","Data":"422ee2aec88f6d9695479975fa9bf7b2c3d3eab9e3d87f14e41463e52aeea3b8"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.474216 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.513640 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.516967 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.517002 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.517018 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.517037 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.517050 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.546687 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.587841 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.620212 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.620247 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.620256 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.620270 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.620280 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.636749 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.673253 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.708302 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.722518 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.722571 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.722586 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.722606 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.722623 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.764283 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.796497 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.824933 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.824969 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.824979 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.824995 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.825007 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.836601 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.868147 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.925803 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.926798 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.926826 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.926835 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.926848 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.926858 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:00Z","lastTransitionTime":"2026-01-24T00:04:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.957704 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:00 crc kubenswrapper[4676]: I0124 00:04:00.989332 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:00Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.028904 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.029813 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.029879 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.029900 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.029926 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.029944 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.073640 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.111827 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.132342 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.132404 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.132416 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.132434 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.132445 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.154355 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.189874 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.211538 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 09:32:18.513892574 +0000 UTC Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.227979 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.235001 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.235056 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.235070 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.235091 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.235103 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.255331 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.255331 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:01 crc kubenswrapper[4676]: E0124 00:04:01.255446 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:01 crc kubenswrapper[4676]: E0124 00:04:01.255511 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.337838 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.337899 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.337919 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.337946 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.337964 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.440676 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.440737 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.440756 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.440780 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.440796 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.475443 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.478560 4676 generic.go:334] "Generic (PLEG): container finished" podID="79ad333b-cf18-4ba3-b9d4-2f89c7c44354" containerID="5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d" exitCode=0 Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.478591 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" event={"ID":"79ad333b-cf18-4ba3-b9d4-2f89c7c44354","Type":"ContainerDied","Data":"5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.493483 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.512873 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.537129 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.546190 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.546256 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.546284 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.546319 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.546332 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.553063 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.566189 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.578870 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.592755 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.604994 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.626088 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.642587 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.648165 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.648187 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.648195 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.648207 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.648216 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.673999 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.708749 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.746108 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.750394 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.750420 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.750429 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.750442 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.750456 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.786017 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.825422 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:01Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.853915 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.853973 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.853994 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.854024 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.854043 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.956492 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.956559 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.956581 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.956612 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:01 crc kubenswrapper[4676]: I0124 00:04:01.956646 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:01Z","lastTransitionTime":"2026-01-24T00:04:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.059444 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.059499 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.059562 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.059591 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.059600 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.161479 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.161515 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.161523 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.161535 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.161544 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.212201 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 14:28:09.969257689 +0000 UTC Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.255663 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:02 crc kubenswrapper[4676]: E0124 00:04:02.255821 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.264216 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.264268 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.264285 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.264307 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.264324 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.367213 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.367271 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.367293 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.367322 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.367342 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.470042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.470082 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.470093 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.470112 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.470124 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.483669 4676 generic.go:334] "Generic (PLEG): container finished" podID="79ad333b-cf18-4ba3-b9d4-2f89c7c44354" containerID="8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239" exitCode=0 Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.483713 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" event={"ID":"79ad333b-cf18-4ba3-b9d4-2f89c7c44354","Type":"ContainerDied","Data":"8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.497670 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.515262 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.530942 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.540659 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.553546 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.566227 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.572559 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.572595 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.572604 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.572621 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.572629 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.578797 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.589274 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.609518 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.624652 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.632757 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.650532 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.666266 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.674079 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.674115 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.674126 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.674144 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.674157 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.679327 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.690730 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:02Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.776739 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.776795 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.776812 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.776834 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.776850 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.880474 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.880514 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.880523 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.880538 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.880547 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.983677 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.983774 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.983791 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.983812 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.983827 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:02Z","lastTransitionTime":"2026-01-24T00:04:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:02 crc kubenswrapper[4676]: I0124 00:04:02.985466 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:04:02 crc kubenswrapper[4676]: E0124 00:04:02.985786 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:10.985739072 +0000 UTC m=+35.015710133 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.085861 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.085894 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.085923 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.085939 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.085940 4676 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.085990 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:11.085973013 +0000 UTC m=+35.115944014 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086020 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086031 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086040 4676 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086064 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:11.086056575 +0000 UTC m=+35.116027576 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086101 4676 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086119 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:11.086113277 +0000 UTC m=+35.116084278 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086156 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086163 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086170 4676 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.086189 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:11.086182149 +0000 UTC m=+35.116153150 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.086955 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.086972 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.086981 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.086992 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.087000 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:03Z","lastTransitionTime":"2026-01-24T00:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.188367 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.188418 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.188431 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.188459 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.188472 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:03Z","lastTransitionTime":"2026-01-24T00:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.212585 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 03:06:44.982662844 +0000 UTC Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.254887 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.254970 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.255026 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:03 crc kubenswrapper[4676]: E0124 00:04:03.255086 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.290341 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.290441 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.290495 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.290520 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.290543 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:03Z","lastTransitionTime":"2026-01-24T00:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.393676 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.393725 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.393740 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.393761 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.393778 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:03Z","lastTransitionTime":"2026-01-24T00:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.492166 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" event={"ID":"79ad333b-cf18-4ba3-b9d4-2f89c7c44354","Type":"ContainerStarted","Data":"c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.496806 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.496833 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.496841 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.496855 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.496864 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:03Z","lastTransitionTime":"2026-01-24T00:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.499786 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.500134 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.513049 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.528300 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.530556 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.547692 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.562996 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.581307 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.599853 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.599922 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.599951 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.599982 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.600004 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:03Z","lastTransitionTime":"2026-01-24T00:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.600786 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.614680 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.634002 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.653363 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.664088 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.678932 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.701669 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.704212 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.704243 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.704252 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.704265 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.704275 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:03Z","lastTransitionTime":"2026-01-24T00:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.722893 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.756437 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.780184 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.806519 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.806568 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.806581 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.806599 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.806611 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:03Z","lastTransitionTime":"2026-01-24T00:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.807248 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.827877 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.843629 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.859680 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.869497 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.881341 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.889821 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.901119 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.909131 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.909163 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.909171 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.909189 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.909200 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:03Z","lastTransitionTime":"2026-01-24T00:04:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.914539 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.928115 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.941110 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.951515 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.960720 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.971310 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:03 crc kubenswrapper[4676]: I0124 00:04:03.981631 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:03Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.011256 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.011317 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.011329 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.011345 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.011355 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.114723 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.114782 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.114801 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.114827 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.114845 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.213339 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 15:02:10.177478845 +0000 UTC Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.217571 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.217613 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.217625 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.217642 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.217653 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.255192 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:04 crc kubenswrapper[4676]: E0124 00:04:04.255460 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.320926 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.320982 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.320994 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.321014 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.321032 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.423975 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.424038 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.424050 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.424073 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.424087 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.502814 4676 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.503371 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.526839 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.526878 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.526892 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.526913 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.526928 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.534095 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.559236 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.581595 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.601894 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.617807 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.629246 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.629289 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.629300 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.629322 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.629334 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.635853 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.646433 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.658666 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.673191 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.689548 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.703557 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.716462 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.728021 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.731961 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.732010 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.732020 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.732042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.732055 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.742359 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.754403 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.774849 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:04Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.834623 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.834694 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.834703 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.834725 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.834740 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.937621 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.937942 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.938042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.938126 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:04 crc kubenswrapper[4676]: I0124 00:04:04.938202 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:04Z","lastTransitionTime":"2026-01-24T00:04:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.041692 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.041766 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.041784 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.041814 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.041828 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.143706 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.143741 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.143749 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.143762 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.143771 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.213666 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 16:26:29.049475747 +0000 UTC Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.246481 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.246518 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.246530 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.246546 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.246559 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.254743 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:05 crc kubenswrapper[4676]: E0124 00:04:05.254843 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.254926 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:05 crc kubenswrapper[4676]: E0124 00:04:05.255138 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.357049 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.357095 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.357105 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.357120 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.357131 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.459105 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.459146 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.459156 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.459171 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.459182 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.508507 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/0.log" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.511943 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886" exitCode=1 Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.511991 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.513126 4676 scope.go:117] "RemoveContainer" containerID="b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.527897 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.539530 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.556772 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:05Z\\\",\\\"message\\\":\\\"\\\\nI0124 00:04:05.114978 5895 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:05.115011 5895 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0124 00:04:05.115016 5895 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0124 00:04:05.115045 5895 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0124 00:04:05.115144 5895 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:05.115169 5895 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:05.115193 5895 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:05.115212 5895 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:05.115222 5895 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0124 00:04:05.115431 5895 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:05.115445 5895 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:05.115449 5895 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:05.115465 5895 factory.go:656] Stopping watch factory\\\\nI0124 00:04:05.115487 5895 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:05.115507 5895 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0124 00:04:05.115523 5895 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.561688 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.561725 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.561736 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.561753 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.561766 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.571401 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.587155 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.598419 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.613364 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.627973 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.647822 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.663957 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.663996 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.664006 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.664022 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.664031 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.665369 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.680576 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.692711 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.704402 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.722008 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.738283 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:05Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.765597 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.765631 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.765641 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.765655 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.765668 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.868136 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.868199 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.868216 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.868240 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.868256 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.970102 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.970136 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.970145 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.970160 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:05 crc kubenswrapper[4676]: I0124 00:04:05.970169 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:05Z","lastTransitionTime":"2026-01-24T00:04:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.072217 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.072296 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.072314 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.072340 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.072358 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:06Z","lastTransitionTime":"2026-01-24T00:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.174640 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.174689 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.174706 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.174732 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.174750 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:06Z","lastTransitionTime":"2026-01-24T00:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.214367 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 05:31:02.700793703 +0000 UTC Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.255272 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:06 crc kubenswrapper[4676]: E0124 00:04:06.255427 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.270739 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.281416 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.281455 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.281467 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.281483 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.281493 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:06Z","lastTransitionTime":"2026-01-24T00:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.283335 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.321785 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:05Z\\\",\\\"message\\\":\\\"\\\\nI0124 00:04:05.114978 5895 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:05.115011 5895 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0124 00:04:05.115016 5895 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0124 00:04:05.115045 5895 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0124 00:04:05.115144 5895 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:05.115169 5895 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:05.115193 5895 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:05.115212 5895 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:05.115222 5895 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0124 00:04:05.115431 5895 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:05.115445 5895 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:05.115449 5895 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:05.115465 5895 factory.go:656] Stopping watch factory\\\\nI0124 00:04:05.115487 5895 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:05.115507 5895 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0124 00:04:05.115523 5895 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.343892 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.360713 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.376043 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.384534 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.384590 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.384603 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.384628 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.384660 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:06Z","lastTransitionTime":"2026-01-24T00:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.392402 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.413631 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.425062 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.443120 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.459294 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.476742 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.486926 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.486974 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.486985 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.487001 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.487011 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:06Z","lastTransitionTime":"2026-01-24T00:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.491481 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.504607 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.517932 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/0.log" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.517919 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.521164 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.521284 4676 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.548270 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.561813 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.577988 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.589457 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.589681 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.589761 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.589851 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.589933 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:06Z","lastTransitionTime":"2026-01-24T00:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.595930 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.615769 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.629336 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.642958 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.679830 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.692843 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.692884 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.692895 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.692913 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.692927 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:06Z","lastTransitionTime":"2026-01-24T00:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.700392 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.719913 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.738540 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.781456 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.796116 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.796166 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.796179 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.796197 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.796211 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:06Z","lastTransitionTime":"2026-01-24T00:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.803676 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.821468 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.856898 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:05Z\\\",\\\"message\\\":\\\"\\\\nI0124 00:04:05.114978 5895 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:05.115011 5895 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0124 00:04:05.115016 5895 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0124 00:04:05.115045 5895 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0124 00:04:05.115144 5895 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:05.115169 5895 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:05.115193 5895 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:05.115212 5895 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:05.115222 5895 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0124 00:04:05.115431 5895 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:05.115445 5895 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:05.115449 5895 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:05.115465 5895 factory.go:656] Stopping watch factory\\\\nI0124 00:04:05.115487 5895 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:05.115507 5895 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0124 00:04:05.115523 5895 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.898368 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.898443 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.898459 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.898481 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:06 crc kubenswrapper[4676]: I0124 00:04:06.898499 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:06Z","lastTransitionTime":"2026-01-24T00:04:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.000905 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.000968 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.000985 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.001009 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.001026 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.103313 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.103421 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.103451 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.103481 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.103498 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.205877 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.205928 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.205945 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.205968 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.205984 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.215018 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 12:24:29.402719641 +0000 UTC Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.254873 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.254873 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:07 crc kubenswrapper[4676]: E0124 00:04:07.255150 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:07 crc kubenswrapper[4676]: E0124 00:04:07.255027 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.309695 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.309762 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.309780 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.309815 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.309833 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.412792 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.412862 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.412886 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.412910 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.412927 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.515814 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.515875 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.515892 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.515918 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.515935 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.526588 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/1.log" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.527289 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/0.log" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.530259 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a" exitCode=1 Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.530312 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.530356 4676 scope.go:117] "RemoveContainer" containerID="b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.531421 4676 scope.go:117] "RemoveContainer" containerID="901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a" Jan 24 00:04:07 crc kubenswrapper[4676]: E0124 00:04:07.531660 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.550862 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.567999 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.598483 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8471192d819a4b266cdc87f77c8ca851671825cfe5b82b786bae9b8f73f1886\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:05Z\\\",\\\"message\\\":\\\"\\\\nI0124 00:04:05.114978 5895 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:05.115011 5895 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0124 00:04:05.115016 5895 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0124 00:04:05.115045 5895 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0124 00:04:05.115144 5895 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:05.115169 5895 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:05.115193 5895 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:05.115212 5895 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:05.115222 5895 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0124 00:04:05.115431 5895 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:05.115445 5895 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:05.115449 5895 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:05.115465 5895 factory.go:656] Stopping watch factory\\\\nI0124 00:04:05.115487 5895 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:05.115507 5895 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0124 00:04:05.115523 5895 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:06Z\\\",\\\"message\\\":\\\"e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0124 00:04:06.463876 6013 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI0124 00:04:06.464703 6013 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI0124 00:04:06.460422 6013 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0124 00:04:06.458980 6013 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0124 00:04:06.464808 6013 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.619575 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.619867 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.620003 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.620125 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.620236 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.621963 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.644084 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.655320 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.678295 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.689227 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.689290 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.689303 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.689322 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.689349 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.693236 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: E0124 00:04:07.706408 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.709234 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.711216 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.711405 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.711441 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.711468 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.711486 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: E0124 00:04:07.727528 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.728801 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.731566 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.731612 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.731628 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.731652 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.731669 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.742504 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: E0124 00:04:07.745266 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.751516 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.751647 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.751734 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.751817 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.751913 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.753101 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.767715 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: E0124 00:04:07.770255 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.773232 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.773259 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.773267 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.773281 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.773302 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.780436 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: E0124 00:04:07.785412 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: E0124 00:04:07.785518 4676 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.786899 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.786958 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.786968 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.786981 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.786990 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.795542 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:07Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.889759 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.890018 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.890128 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.890305 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.890424 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.993908 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.993974 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.993997 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.994028 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:07 crc kubenswrapper[4676]: I0124 00:04:07.994053 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:07Z","lastTransitionTime":"2026-01-24T00:04:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.097661 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.097714 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.097728 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.097750 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.097765 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:08Z","lastTransitionTime":"2026-01-24T00:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.199897 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.199938 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.199953 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.199970 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.199981 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:08Z","lastTransitionTime":"2026-01-24T00:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.215555 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 03:57:35.449626457 +0000 UTC Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.255174 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:08 crc kubenswrapper[4676]: E0124 00:04:08.255444 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.302923 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.302987 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.303004 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.303027 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.303059 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:08Z","lastTransitionTime":"2026-01-24T00:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.406068 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.406140 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.406166 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.406195 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.406216 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:08Z","lastTransitionTime":"2026-01-24T00:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.509015 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.509055 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.509066 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.509082 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.509095 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:08Z","lastTransitionTime":"2026-01-24T00:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.535125 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/1.log" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.538294 4676 scope.go:117] "RemoveContainer" containerID="901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a" Jan 24 00:04:08 crc kubenswrapper[4676]: E0124 00:04:08.538446 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.553078 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.565685 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.576677 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.590399 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.610935 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.610965 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.610973 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.611002 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.611012 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:08Z","lastTransitionTime":"2026-01-24T00:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.658514 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.669974 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.683921 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.695122 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.712877 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.712931 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.713046 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.713098 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.713117 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:08Z","lastTransitionTime":"2026-01-24T00:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.716978 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:06Z\\\",\\\"message\\\":\\\"e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0124 00:04:06.463876 6013 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI0124 00:04:06.464703 6013 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI0124 00:04:06.460422 6013 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0124 00:04:06.458980 6013 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0124 00:04:06.464808 6013 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.737122 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.753799 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.765647 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.776996 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.788546 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.797049 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:08Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.815700 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.815724 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.815732 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.815745 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.815769 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:08Z","lastTransitionTime":"2026-01-24T00:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.918331 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.918410 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.918427 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.918448 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:08 crc kubenswrapper[4676]: I0124 00:04:08.918463 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:08Z","lastTransitionTime":"2026-01-24T00:04:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.021425 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.021473 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.021488 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.021507 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.021523 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.124591 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.124664 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.124682 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.124708 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.124727 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.216218 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 01:12:35.154406132 +0000 UTC Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.227652 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.227717 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.227732 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.227751 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.227766 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.254356 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts"] Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.255199 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.255248 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: E0124 00:04:09.255444 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.255487 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:09 crc kubenswrapper[4676]: E0124 00:04:09.255749 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.259102 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.262298 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.278405 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.297299 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.315102 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.330814 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.330845 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.330855 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.330868 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.330878 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.334798 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.345834 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sj4f8\" (UniqueName: \"kubernetes.io/projected/40151406-46c7-4668-8b2b-db0585847be9-kube-api-access-sj4f8\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.345909 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40151406-46c7-4668-8b2b-db0585847be9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.345975 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/40151406-46c7-4668-8b2b-db0585847be9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.346003 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/40151406-46c7-4668-8b2b-db0585847be9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.352726 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.365415 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.382207 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.399531 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.421715 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.433743 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.434007 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.434081 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.434203 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.434274 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.439760 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.446661 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/40151406-46c7-4668-8b2b-db0585847be9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.446708 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/40151406-46c7-4668-8b2b-db0585847be9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.446735 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sj4f8\" (UniqueName: \"kubernetes.io/projected/40151406-46c7-4668-8b2b-db0585847be9-kube-api-access-sj4f8\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.446780 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40151406-46c7-4668-8b2b-db0585847be9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.447677 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/40151406-46c7-4668-8b2b-db0585847be9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.447757 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/40151406-46c7-4668-8b2b-db0585847be9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.457259 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/40151406-46c7-4668-8b2b-db0585847be9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.459200 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.463796 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sj4f8\" (UniqueName: \"kubernetes.io/projected/40151406-46c7-4668-8b2b-db0585847be9-kube-api-access-sj4f8\") pod \"ovnkube-control-plane-749d76644c-7m8ts\" (UID: \"40151406-46c7-4668-8b2b-db0585847be9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.473902 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.487004 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.499542 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.527540 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:06Z\\\",\\\"message\\\":\\\"e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0124 00:04:06.463876 6013 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI0124 00:04:06.464703 6013 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI0124 00:04:06.460422 6013 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0124 00:04:06.458980 6013 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0124 00:04:06.464808 6013 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.536467 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.536533 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.536555 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.536586 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.536607 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.559313 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.569592 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" Jan 24 00:04:09 crc kubenswrapper[4676]: W0124 00:04:09.587802 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40151406_46c7_4668_8b2b_db0585847be9.slice/crio-0ac6ade2d0240e340fe022a220fd0929fae2090fb4532fd0c2cd300e5c65aa22 WatchSource:0}: Error finding container 0ac6ade2d0240e340fe022a220fd0929fae2090fb4532fd0c2cd300e5c65aa22: Status 404 returned error can't find the container with id 0ac6ade2d0240e340fe022a220fd0929fae2090fb4532fd0c2cd300e5c65aa22 Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.639010 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.639046 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.639058 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.639074 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.639086 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.741834 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.741874 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.741886 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.741902 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.741915 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.844112 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.844370 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.844393 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.844409 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.844420 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.947148 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.947193 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.947209 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.947232 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:09 crc kubenswrapper[4676]: I0124 00:04:09.947249 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:09Z","lastTransitionTime":"2026-01-24T00:04:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.049940 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.050167 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.050260 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.050350 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.050448 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.153239 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.153279 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.153288 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.153304 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.153314 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.216996 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 13:25:07.113228846 +0000 UTC Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.254606 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:10 crc kubenswrapper[4676]: E0124 00:04:10.254696 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.255048 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.255136 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.255200 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.255301 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.255371 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.357635 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.357690 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.357703 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.357723 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.357739 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.402368 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-r4q22"] Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.403090 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:10 crc kubenswrapper[4676]: E0124 00:04:10.403186 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.420589 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.438981 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.460686 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.461007 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsw85\" (UniqueName: \"kubernetes.io/projected/18335446-e572-4741-ad9e-e7aadee7550b-kube-api-access-tsw85\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.461280 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.462821 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.462875 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.462887 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.462907 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.462919 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.484171 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:06Z\\\",\\\"message\\\":\\\"e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0124 00:04:06.463876 6013 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI0124 00:04:06.464703 6013 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI0124 00:04:06.460422 6013 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0124 00:04:06.458980 6013 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0124 00:04:06.464808 6013 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.502438 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.517214 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.530401 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.547124 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.547404 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" event={"ID":"40151406-46c7-4668-8b2b-db0585847be9","Type":"ContainerStarted","Data":"7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.547451 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" event={"ID":"40151406-46c7-4668-8b2b-db0585847be9","Type":"ContainerStarted","Data":"4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.547465 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" event={"ID":"40151406-46c7-4668-8b2b-db0585847be9","Type":"ContainerStarted","Data":"0ac6ade2d0240e340fe022a220fd0929fae2090fb4532fd0c2cd300e5c65aa22"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.561346 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.562279 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsw85\" (UniqueName: \"kubernetes.io/projected/18335446-e572-4741-ad9e-e7aadee7550b-kube-api-access-tsw85\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.562572 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:10 crc kubenswrapper[4676]: E0124 00:04:10.562782 4676 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:10 crc kubenswrapper[4676]: E0124 00:04:10.562924 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs podName:18335446-e572-4741-ad9e-e7aadee7550b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:11.062891328 +0000 UTC m=+35.092862369 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs") pod "network-metrics-daemon-r4q22" (UID: "18335446-e572-4741-ad9e-e7aadee7550b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.565540 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.565603 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.565626 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.565657 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.565679 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.589596 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsw85\" (UniqueName: \"kubernetes.io/projected/18335446-e572-4741-ad9e-e7aadee7550b-kube-api-access-tsw85\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.595196 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.609708 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.625464 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.643747 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.660649 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.668779 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.668811 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.668825 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.668844 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.668857 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.676290 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.691566 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.705204 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.723212 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.754841 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:06Z\\\",\\\"message\\\":\\\"e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0124 00:04:06.463876 6013 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI0124 00:04:06.464703 6013 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI0124 00:04:06.460422 6013 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0124 00:04:06.458980 6013 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0124 00:04:06.464808 6013 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.770916 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.770971 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.770980 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.770998 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.771007 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.773270 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.786603 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.799430 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.815047 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.822090 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.840978 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.864473 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.873254 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.873426 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.873519 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.873663 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.873803 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.881420 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.896168 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.911726 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.931542 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.950886 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.969006 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.975757 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.975869 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.975967 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.976060 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.976135 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:10Z","lastTransitionTime":"2026-01-24T00:04:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:10 crc kubenswrapper[4676]: I0124 00:04:10.983686 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:10.999897 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:10Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.018196 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.034874 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.063764 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:06Z\\\",\\\"message\\\":\\\"e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0124 00:04:06.463876 6013 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI0124 00:04:06.464703 6013 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI0124 00:04:06.460422 6013 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0124 00:04:06.458980 6013 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0124 00:04:06.464808 6013 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.066845 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.067119 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:27.067078773 +0000 UTC m=+51.097049814 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.067208 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.067428 4676 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.067522 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs podName:18335446-e572-4741-ad9e-e7aadee7550b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:12.067497605 +0000 UTC m=+36.097468646 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs") pod "network-metrics-daemon-r4q22" (UID: "18335446-e572-4741-ad9e-e7aadee7550b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.078668 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.078731 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.078755 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.078786 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.078811 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:11Z","lastTransitionTime":"2026-01-24T00:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.086117 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.106075 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.126644 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.155049 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.168209 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.168274 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.168331 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.168365 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168426 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168457 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168474 4676 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168502 4676 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168538 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:27.16851717 +0000 UTC m=+51.198488191 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168605 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168619 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168630 4676 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168662 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:27.168651054 +0000 UTC m=+51.198622065 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168698 4676 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168721 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:27.168713686 +0000 UTC m=+51.198684697 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.168739 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:27.168733016 +0000 UTC m=+51.198704027 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.169112 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.181177 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.181215 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.181228 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.181245 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.181257 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:11Z","lastTransitionTime":"2026-01-24T00:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.200701 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.217884 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.218178 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 05:29:40.747970379 +0000 UTC Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.238014 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.255509 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.255552 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.255523 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.255679 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:11 crc kubenswrapper[4676]: E0124 00:04:11.255730 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.270985 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.283513 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.283559 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.283575 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.283596 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.283614 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:11Z","lastTransitionTime":"2026-01-24T00:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.288509 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.301012 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.315937 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.331468 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.348534 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:11Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.385813 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.385854 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.385865 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.385879 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.385891 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:11Z","lastTransitionTime":"2026-01-24T00:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.488255 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.488299 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.488307 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.488321 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.488331 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:11Z","lastTransitionTime":"2026-01-24T00:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.591033 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.591085 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.591101 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.591120 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.591131 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:11Z","lastTransitionTime":"2026-01-24T00:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.694211 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.694265 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.694277 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.694296 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.694310 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:11Z","lastTransitionTime":"2026-01-24T00:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.796712 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.796750 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.796759 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.796772 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.796782 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:11Z","lastTransitionTime":"2026-01-24T00:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.899867 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.899941 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.899979 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.900011 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:11 crc kubenswrapper[4676]: I0124 00:04:11.900034 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:11Z","lastTransitionTime":"2026-01-24T00:04:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.003250 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.003301 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.003318 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.003343 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.003359 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.078047 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:12 crc kubenswrapper[4676]: E0124 00:04:12.078186 4676 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:12 crc kubenswrapper[4676]: E0124 00:04:12.078240 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs podName:18335446-e572-4741-ad9e-e7aadee7550b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:14.078223665 +0000 UTC m=+38.108194676 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs") pod "network-metrics-daemon-r4q22" (UID: "18335446-e572-4741-ad9e-e7aadee7550b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.106551 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.106616 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.106629 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.106653 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.106670 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.210294 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.210363 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.210427 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.210452 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.210466 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.218909 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 08:44:53.662524312 +0000 UTC Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.255637 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.255819 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:12 crc kubenswrapper[4676]: E0124 00:04:12.255962 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:12 crc kubenswrapper[4676]: E0124 00:04:12.256206 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.312967 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.313029 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.313044 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.313070 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.313085 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.416674 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.416745 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.416757 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.416777 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.416795 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.525879 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.525944 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.525961 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.525987 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.526008 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.629234 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.629297 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.629317 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.629361 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.629420 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.732104 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.732157 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.732173 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.732199 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.732215 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.836002 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.836080 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.836103 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.836133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.836156 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.938677 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.938748 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.938772 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.938796 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:12 crc kubenswrapper[4676]: I0124 00:04:12.938814 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:12Z","lastTransitionTime":"2026-01-24T00:04:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.041372 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.041500 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.041524 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.041557 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.041580 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.144776 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.144891 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.144915 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.144945 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.144967 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.219987 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 05:53:17.713937971 +0000 UTC Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.248222 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.248272 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.248294 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.248322 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.248348 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.254892 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.254895 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:13 crc kubenswrapper[4676]: E0124 00:04:13.255035 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:13 crc kubenswrapper[4676]: E0124 00:04:13.255141 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.351133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.351187 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.351223 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.351243 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.351258 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.454044 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.454109 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.454123 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.454140 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.454153 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.557894 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.557974 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.558001 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.558031 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.558053 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.660446 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.660507 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.660518 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.660537 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.660549 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.763168 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.763230 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.763245 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.763262 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.763274 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.865933 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.865975 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.865997 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.866015 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.866027 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.968768 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.968826 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.968843 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.968869 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:13 crc kubenswrapper[4676]: I0124 00:04:13.968886 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:13Z","lastTransitionTime":"2026-01-24T00:04:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.071709 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.071766 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.071788 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.071817 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.071842 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.100759 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:14 crc kubenswrapper[4676]: E0124 00:04:14.100940 4676 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:14 crc kubenswrapper[4676]: E0124 00:04:14.100998 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs podName:18335446-e572-4741-ad9e-e7aadee7550b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:18.100983803 +0000 UTC m=+42.130954804 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs") pod "network-metrics-daemon-r4q22" (UID: "18335446-e572-4741-ad9e-e7aadee7550b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.174846 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.174906 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.174923 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.174947 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.174965 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.220668 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 16:36:40.069944354 +0000 UTC Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.254991 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.255059 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:14 crc kubenswrapper[4676]: E0124 00:04:14.255193 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:14 crc kubenswrapper[4676]: E0124 00:04:14.255336 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.277198 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.277255 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.277273 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.277296 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.277339 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.379890 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.379966 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.379984 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.380011 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.380030 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.482434 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.482469 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.482476 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.482488 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.482498 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.585143 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.585177 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.585187 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.585202 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.585212 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.688495 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.688577 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.688611 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.688643 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.688662 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.791219 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.791305 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.791324 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.791370 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.791408 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.894418 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.894479 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.894497 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.894522 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.894539 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.997579 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.997642 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.997660 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.997689 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:14 crc kubenswrapper[4676]: I0124 00:04:14.997714 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:14Z","lastTransitionTime":"2026-01-24T00:04:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.101868 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.101943 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.101962 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.101988 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.102007 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:15Z","lastTransitionTime":"2026-01-24T00:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.204034 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.204077 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.204087 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.204103 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.204114 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:15Z","lastTransitionTime":"2026-01-24T00:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.221493 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 03:24:37.995985575 +0000 UTC Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.254840 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.254858 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:15 crc kubenswrapper[4676]: E0124 00:04:15.255023 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:15 crc kubenswrapper[4676]: E0124 00:04:15.255124 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.307353 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.307481 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.307507 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.307539 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.307560 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:15Z","lastTransitionTime":"2026-01-24T00:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.411821 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.411877 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.411894 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.411919 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.411938 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:15Z","lastTransitionTime":"2026-01-24T00:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.515579 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.515632 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.515651 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.515677 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.515696 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:15Z","lastTransitionTime":"2026-01-24T00:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.619142 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.619221 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.619319 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.619411 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.619440 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:15Z","lastTransitionTime":"2026-01-24T00:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.722649 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.722685 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.722697 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.722713 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.722724 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:15Z","lastTransitionTime":"2026-01-24T00:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.825771 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.825829 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.825846 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.825870 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.825891 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:15Z","lastTransitionTime":"2026-01-24T00:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.928677 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.928751 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.928767 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.928788 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:15 crc kubenswrapper[4676]: I0124 00:04:15.928831 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:15Z","lastTransitionTime":"2026-01-24T00:04:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.031605 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.031655 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.031701 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.031724 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.031740 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.135498 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.135553 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.135575 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.135601 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.135618 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.222401 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 13:10:50.34454338 +0000 UTC Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.238927 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.238971 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.238983 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.239005 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.239017 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.255507 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.255523 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:16 crc kubenswrapper[4676]: E0124 00:04:16.255647 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:16 crc kubenswrapper[4676]: E0124 00:04:16.255909 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.270997 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.287659 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.308983 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.333979 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.342476 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.342539 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.342559 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.342584 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.342602 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.352339 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.369321 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.385944 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.401581 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.417818 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.444059 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:06Z\\\",\\\"message\\\":\\\"e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0124 00:04:06.463876 6013 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI0124 00:04:06.464703 6013 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI0124 00:04:06.460422 6013 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0124 00:04:06.458980 6013 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0124 00:04:06.464808 6013 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.445848 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.445923 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.445941 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.445966 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.445985 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.461486 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.494261 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.510000 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.523527 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.539288 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.548167 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.548194 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.548202 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.548216 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.548225 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.558190 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.567907 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:16Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.650718 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.650961 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.650971 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.650986 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.650995 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.753893 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.753988 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.754010 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.754066 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.754084 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.856771 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.856845 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.856862 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.856887 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.856908 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.959804 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.959858 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.959873 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.959897 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:16 crc kubenswrapper[4676]: I0124 00:04:16.959919 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:16Z","lastTransitionTime":"2026-01-24T00:04:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.062915 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.062989 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.063011 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.063039 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.063060 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.165734 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.165806 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.165829 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.165858 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.165879 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.222885 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 04:46:59.064603595 +0000 UTC Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.255504 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:17 crc kubenswrapper[4676]: E0124 00:04:17.256098 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.255548 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:17 crc kubenswrapper[4676]: E0124 00:04:17.256290 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.270042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.270121 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.270143 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.270174 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.270193 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.373365 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.373472 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.373493 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.373518 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.373539 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.476316 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.476411 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.476425 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.476449 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.476462 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.579174 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.579235 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.579255 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.579320 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.579341 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.683021 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.683093 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.683115 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.683145 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.683162 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.786982 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.787060 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.787079 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.787105 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.787125 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.890252 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.890313 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.890331 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.890354 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.890419 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.993289 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.993471 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.993499 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.993536 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:17 crc kubenswrapper[4676]: I0124 00:04:17.993562 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:17Z","lastTransitionTime":"2026-01-24T00:04:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.068108 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.068156 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.068171 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.068188 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.068200 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.085297 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:18Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.090347 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.090408 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.090423 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.090445 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.090458 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.101934 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:18Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.105974 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.106019 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.106032 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.106052 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.106063 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.119402 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:18Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.123691 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.123737 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.123750 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.123770 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.123784 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.135138 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:18Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.138684 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.138712 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.138721 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.138743 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.138755 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.146216 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.146442 4676 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.146546 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs podName:18335446-e572-4741-ad9e-e7aadee7550b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:26.14652001 +0000 UTC m=+50.176491011 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs") pod "network-metrics-daemon-r4q22" (UID: "18335446-e572-4741-ad9e-e7aadee7550b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.149570 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:18Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.149726 4676 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.151330 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.151359 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.151370 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.151403 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.151417 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.224047 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 01:59:43.370394791 +0000 UTC Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.254558 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.254625 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.254648 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.254675 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.254696 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.254885 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.254976 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.255100 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:18 crc kubenswrapper[4676]: E0124 00:04:18.255216 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.357028 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.357074 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.357088 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.357108 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.357121 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.460117 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.460191 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.460216 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.460240 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.460264 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.563015 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.563088 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.563110 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.563138 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.563159 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.666066 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.666202 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.666227 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.666254 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.666274 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.770107 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.770252 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.770280 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.770342 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.770429 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.873910 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.873963 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.874008 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.874031 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.874049 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.976603 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.976646 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.976654 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.976668 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:18 crc kubenswrapper[4676]: I0124 00:04:18.976679 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:18Z","lastTransitionTime":"2026-01-24T00:04:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.079145 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.079175 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.079185 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.079198 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.079208 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:19Z","lastTransitionTime":"2026-01-24T00:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.178100 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.179422 4676 scope.go:117] "RemoveContainer" containerID="901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.181867 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.181898 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.181910 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.181927 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.181939 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:19Z","lastTransitionTime":"2026-01-24T00:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.224752 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 18:29:49.248811921 +0000 UTC Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.254635 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.254660 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:19 crc kubenswrapper[4676]: E0124 00:04:19.254752 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:19 crc kubenswrapper[4676]: E0124 00:04:19.254869 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.284609 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.284654 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.284671 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.284693 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.284710 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:19Z","lastTransitionTime":"2026-01-24T00:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.388147 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.388197 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.388216 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.388242 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.388259 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:19Z","lastTransitionTime":"2026-01-24T00:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.491206 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.491269 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.491282 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.491308 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.491324 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:19Z","lastTransitionTime":"2026-01-24T00:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.581897 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/1.log" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.584937 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.593184 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.593217 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.593227 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.593243 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.593254 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:19Z","lastTransitionTime":"2026-01-24T00:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.696286 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.696317 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.696325 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.696339 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.696348 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:19Z","lastTransitionTime":"2026-01-24T00:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.799866 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.799976 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.800053 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.800132 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.800165 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:19Z","lastTransitionTime":"2026-01-24T00:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.903139 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.903181 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.903196 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.903214 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:19 crc kubenswrapper[4676]: I0124 00:04:19.903231 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:19Z","lastTransitionTime":"2026-01-24T00:04:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.005510 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.005542 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.005551 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.005564 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.005574 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.107783 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.107812 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.107820 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.107833 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.107842 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.209989 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.210020 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.210028 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.210042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.210050 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.225614 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 11:16:08.18041271 +0000 UTC Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.255021 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.255032 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:20 crc kubenswrapper[4676]: E0124 00:04:20.255163 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:20 crc kubenswrapper[4676]: E0124 00:04:20.255571 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.312539 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.312571 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.312581 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.312600 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.312614 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.415680 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.415717 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.415732 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.415753 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.415768 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.518115 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.518164 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.518182 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.518206 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.518226 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.589556 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.610291 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.621278 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.621332 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.621351 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.621402 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.621422 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.629131 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.657162 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:06Z\\\",\\\"message\\\":\\\"e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0124 00:04:06.463876 6013 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI0124 00:04:06.464703 6013 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI0124 00:04:06.460422 6013 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0124 00:04:06.458980 6013 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0124 00:04:06.464808 6013 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.675684 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.685831 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.697339 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.707200 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.723871 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.723904 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.723912 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.723927 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.723936 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.728194 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.740629 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.752050 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.761937 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.778455 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.790063 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.804926 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.819620 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.825976 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.826025 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.826038 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.826057 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.826069 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.835162 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.846468 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:20Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.928459 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.928492 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.928503 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.928519 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:20 crc kubenswrapper[4676]: I0124 00:04:20.928529 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:20Z","lastTransitionTime":"2026-01-24T00:04:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.034654 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.034723 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.034741 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.034768 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.034793 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.138053 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.138109 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.138127 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.138152 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.138169 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.226335 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 21:40:59.247113713 +0000 UTC Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.240708 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.240759 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.240776 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.240800 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.240817 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.255661 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.255728 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:21 crc kubenswrapper[4676]: E0124 00:04:21.255827 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:21 crc kubenswrapper[4676]: E0124 00:04:21.256052 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.344343 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.344487 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.344509 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.344539 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.344560 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.447822 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.447902 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.447923 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.447954 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.447973 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.550803 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.550860 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.550872 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.550891 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.550904 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.595651 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/2.log" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.596782 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/1.log" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.601128 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76" exitCode=1 Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.601196 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.601270 4676 scope.go:117] "RemoveContainer" containerID="901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.602504 4676 scope.go:117] "RemoveContainer" containerID="a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76" Jan 24 00:04:21 crc kubenswrapper[4676]: E0124 00:04:21.602885 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.624787 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.643958 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.652963 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.653032 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.653049 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.653069 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.653083 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.660685 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.676587 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.694597 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.711018 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.723428 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.735356 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.744725 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.755073 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.755358 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.755428 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.755453 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.755471 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.763896 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://901a2ef8ebf68032c2bbcb3d41f57a1ff938b482904ce7cbbb377f04418ff15a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:06Z\\\",\\\"message\\\":\\\"e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0124 00:04:06.463876 6013 ovnkube_controller.go:900] Cache entry expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" but failed to find it\\\\nI0124 00:04:06.464703 6013 ovnkube_controller.go:804] Add Logical Switch Port event expected pod with UID \\\\\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\\\\\" in cache\\\\nI0124 00:04:06.460422 6013 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0124 00:04:06.458980 6013 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-dns/dns-default]} name:Service_openshift-dns/dns-default_TCP_node_router+switch_crc options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.10:53: 10.217.4.10:9154:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {be9dcc9e-c16a-4962-a6d2-4adeb0b929c4}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0124 00:04:06.464808 6013 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.775480 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.800473 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.816480 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.836984 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.847894 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.856940 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.856962 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.856969 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.856981 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.856990 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.861161 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.874359 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:21Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.959737 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.959765 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.959774 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.959785 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:21 crc kubenswrapper[4676]: I0124 00:04:21.959792 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:21Z","lastTransitionTime":"2026-01-24T00:04:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.062594 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.062653 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.062670 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.062692 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.062708 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.165858 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.165958 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.165974 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.165998 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.166017 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.227132 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 02:26:51.20875663 +0000 UTC Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.255711 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.255761 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:22 crc kubenswrapper[4676]: E0124 00:04:22.255879 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:22 crc kubenswrapper[4676]: E0124 00:04:22.256006 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.268057 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.268105 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.268123 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.268150 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.268173 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.371573 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.371642 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.371659 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.371690 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.371714 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.474282 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.474367 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.474430 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.474466 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.474489 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.576904 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.576971 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.576988 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.577011 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.577028 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.606766 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/2.log" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.612114 4676 scope.go:117] "RemoveContainer" containerID="a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76" Jan 24 00:04:22 crc kubenswrapper[4676]: E0124 00:04:22.612345 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.634431 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.651372 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.666520 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.680463 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.680487 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.680494 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.680507 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.680515 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.684834 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.701047 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.716101 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.729133 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.741931 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.755416 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.770648 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.783058 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.783117 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.783134 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.783192 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.783212 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.789420 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.807453 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.823484 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.837498 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.860088 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.874646 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.885313 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.885343 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.885352 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.885367 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.885392 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.903225 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:22Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.987820 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.987860 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.987871 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.987889 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:22 crc kubenswrapper[4676]: I0124 00:04:22.987900 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:22Z","lastTransitionTime":"2026-01-24T00:04:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.091182 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.091252 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.091278 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.091311 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.091334 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:23Z","lastTransitionTime":"2026-01-24T00:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.193711 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.193773 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.193797 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.193825 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.193847 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:23Z","lastTransitionTime":"2026-01-24T00:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.227714 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 08:56:32.301149403 +0000 UTC Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.255422 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.255433 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:23 crc kubenswrapper[4676]: E0124 00:04:23.255635 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:23 crc kubenswrapper[4676]: E0124 00:04:23.255741 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.296739 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.296801 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.296823 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.296854 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.296877 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:23Z","lastTransitionTime":"2026-01-24T00:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.400005 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.400071 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.400088 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.400113 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.400131 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:23Z","lastTransitionTime":"2026-01-24T00:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.503192 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.503257 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.503269 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.503291 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.503302 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:23Z","lastTransitionTime":"2026-01-24T00:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.605852 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.605911 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.605925 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.605946 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.605960 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:23Z","lastTransitionTime":"2026-01-24T00:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.708189 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.708311 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.708334 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.708465 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.708483 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:23Z","lastTransitionTime":"2026-01-24T00:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.810836 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.810869 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.810877 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.810889 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.810901 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:23Z","lastTransitionTime":"2026-01-24T00:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.913941 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.914007 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.914025 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.914052 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:23 crc kubenswrapper[4676]: I0124 00:04:23.914071 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:23Z","lastTransitionTime":"2026-01-24T00:04:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.017354 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.017435 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.017454 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.017478 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.017495 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.120789 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.120844 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.120863 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.120888 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.120910 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.193778 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.206709 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.212701 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.223942 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.224058 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.224083 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.224149 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.224180 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.228272 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 05:37:50.3642817 +0000 UTC Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.233118 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.251629 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.255163 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.255163 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:24 crc kubenswrapper[4676]: E0124 00:04:24.255350 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:24 crc kubenswrapper[4676]: E0124 00:04:24.255516 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.272898 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.293568 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.315159 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.327027 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.327154 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.327173 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.327198 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.327217 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.331350 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.350291 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.367216 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.399678 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.418968 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.429923 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.430013 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.430054 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.430090 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.430111 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.445949 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.461062 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.483697 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.503235 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.520814 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.532684 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.532712 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.532720 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.532735 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.532745 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.536413 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:24Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.634706 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.634770 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.634787 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.634811 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.634834 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.737234 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.737289 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.737311 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.737336 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.737365 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.840059 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.840128 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.840151 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.840183 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.840207 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.943506 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.943564 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.943582 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.943611 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:24 crc kubenswrapper[4676]: I0124 00:04:24.943632 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:24Z","lastTransitionTime":"2026-01-24T00:04:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.046405 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.046445 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.046455 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.046471 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.046481 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.149419 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.149835 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.149927 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.150021 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.150109 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.228407 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 04:44:12.006115868 +0000 UTC Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.253046 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.253113 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.253130 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.253160 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.253180 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.255426 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:25 crc kubenswrapper[4676]: E0124 00:04:25.255597 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.255427 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:25 crc kubenswrapper[4676]: E0124 00:04:25.256165 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.356170 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.356232 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.356250 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.356276 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.356312 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.459306 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.459353 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.459372 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.459441 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.459458 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.562514 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.562571 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.562592 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.562626 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.562650 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.665927 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.665981 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.665997 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.666021 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.666039 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.769146 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.769214 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.769233 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.769258 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.769275 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.872597 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.872694 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.872717 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.872744 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.872764 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.975678 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.975749 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.975767 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.975792 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:25 crc kubenswrapper[4676]: I0124 00:04:25.975810 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:25Z","lastTransitionTime":"2026-01-24T00:04:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.079072 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.079148 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.079171 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.079203 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.079225 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:26Z","lastTransitionTime":"2026-01-24T00:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.182177 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.182248 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.182265 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.182292 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.182311 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:26Z","lastTransitionTime":"2026-01-24T00:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.229044 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 10:42:26.9757949 +0000 UTC Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.235779 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:26 crc kubenswrapper[4676]: E0124 00:04:26.236047 4676 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:26 crc kubenswrapper[4676]: E0124 00:04:26.236138 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs podName:18335446-e572-4741-ad9e-e7aadee7550b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:42.236107119 +0000 UTC m=+66.266078150 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs") pod "network-metrics-daemon-r4q22" (UID: "18335446-e572-4741-ad9e-e7aadee7550b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.254842 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.255247 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:26 crc kubenswrapper[4676]: E0124 00:04:26.255629 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:26 crc kubenswrapper[4676]: E0124 00:04:26.256135 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.285581 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.285643 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.285700 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.285724 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.285740 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:26Z","lastTransitionTime":"2026-01-24T00:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.289963 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.308705 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.327725 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.348019 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.376885 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.389424 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.389474 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.389492 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.389516 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.389533 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:26Z","lastTransitionTime":"2026-01-24T00:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.393802 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.424093 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.443804 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.462413 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.480066 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.492192 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.492258 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.492283 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.492311 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.492333 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:26Z","lastTransitionTime":"2026-01-24T00:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.500795 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.521204 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.544452 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.561485 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.583265 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.594313 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.594428 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.594453 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.594478 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.594495 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:26Z","lastTransitionTime":"2026-01-24T00:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.599186 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.618997 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.635396 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:26Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.697179 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.697217 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.697226 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.697241 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.697252 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:26Z","lastTransitionTime":"2026-01-24T00:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.799649 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.800006 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.800020 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.800038 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.800053 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:26Z","lastTransitionTime":"2026-01-24T00:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.902425 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.902468 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.902480 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.902496 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:26 crc kubenswrapper[4676]: I0124 00:04:26.902509 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:26Z","lastTransitionTime":"2026-01-24T00:04:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.006171 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.006218 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.006230 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.006246 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.006257 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.109448 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.109554 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.109644 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.109709 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.109725 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.145422 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.145672 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:04:59.145634419 +0000 UTC m=+83.175605470 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.213358 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.213448 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.213467 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.213492 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.213511 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.229794 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 00:15:22.811558955 +0000 UTC Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.247278 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.247346 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.247434 4676 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.247520 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:59.247495749 +0000 UTC m=+83.277466790 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.247565 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.247589 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.247606 4676 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.247664 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:59.247645193 +0000 UTC m=+83.277616234 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.247446 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.248294 4676 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.248502 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:59.248444857 +0000 UTC m=+83.278415898 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.247718 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.248625 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.248672 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.248702 4676 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.248783 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-24 00:04:59.248760636 +0000 UTC m=+83.278731667 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.255555 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.255570 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.255774 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:27 crc kubenswrapper[4676]: E0124 00:04:27.256105 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.317654 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.317828 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.317863 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.317894 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.317925 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.420960 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.421349 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.421528 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.421692 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.421881 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.524773 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.524819 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.524834 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.524854 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.524868 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.627783 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.627827 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.627843 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.627864 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.627879 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.730011 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.730264 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.730352 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.730451 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.730513 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.833279 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.833334 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.833348 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.833369 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.833404 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.935070 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.935289 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.935347 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.935430 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:27 crc kubenswrapper[4676]: I0124 00:04:27.935501 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:27Z","lastTransitionTime":"2026-01-24T00:04:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.038080 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.038115 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.038128 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.038144 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.038153 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.141296 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.141369 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.141437 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.141467 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.141489 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.230755 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 05:51:54.46834656 +0000 UTC Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.233129 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.233362 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.233570 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.233709 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.233851 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.254906 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.254992 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:28 crc kubenswrapper[4676]: E0124 00:04:28.255046 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:28 crc kubenswrapper[4676]: E0124 00:04:28.255177 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:28 crc kubenswrapper[4676]: E0124 00:04:28.256074 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:28Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.261210 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.261240 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.261250 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.261265 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.261275 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: E0124 00:04:28.280619 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:28Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.286197 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.286253 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.286272 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.286296 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.286314 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: E0124 00:04:28.303650 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:28Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.308726 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.308830 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.308847 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.308895 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.308916 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: E0124 00:04:28.329187 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:28Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.335052 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.335114 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.335134 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.335159 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.335181 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: E0124 00:04:28.352685 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:28Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:28 crc kubenswrapper[4676]: E0124 00:04:28.352932 4676 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.354979 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.355043 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.355062 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.355086 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.355104 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.457630 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.457689 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.457704 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.457726 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.457740 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.560691 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.560732 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.560748 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.560767 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.560782 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.663513 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.663570 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.663589 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.663616 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.663633 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.769457 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.769514 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.769531 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.769555 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.769572 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.875280 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.875325 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.875339 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.875356 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.875389 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.977705 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.977778 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.977796 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.977823 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:28 crc kubenswrapper[4676]: I0124 00:04:28.977841 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:28Z","lastTransitionTime":"2026-01-24T00:04:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.080090 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.080212 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.080240 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.080272 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.080297 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:29Z","lastTransitionTime":"2026-01-24T00:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.183180 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.183550 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.183683 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.183789 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.183870 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:29Z","lastTransitionTime":"2026-01-24T00:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.231447 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 17:36:41.61309151 +0000 UTC Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.254767 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.254767 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:29 crc kubenswrapper[4676]: E0124 00:04:29.255428 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:29 crc kubenswrapper[4676]: E0124 00:04:29.255889 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.286532 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.286572 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.286582 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.286596 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.286607 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:29Z","lastTransitionTime":"2026-01-24T00:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.389413 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.389450 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.389462 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.389478 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.389490 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:29Z","lastTransitionTime":"2026-01-24T00:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.492095 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.492497 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.492581 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.492654 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.492721 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:29Z","lastTransitionTime":"2026-01-24T00:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.595387 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.595432 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.595443 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.595459 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.595473 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:29Z","lastTransitionTime":"2026-01-24T00:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.698080 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.698148 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.698170 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.698200 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.698217 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:29Z","lastTransitionTime":"2026-01-24T00:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.800902 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.800969 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.800993 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.801020 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.801041 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:29Z","lastTransitionTime":"2026-01-24T00:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.904768 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.904828 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.904847 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.904872 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:29 crc kubenswrapper[4676]: I0124 00:04:29.904891 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:29Z","lastTransitionTime":"2026-01-24T00:04:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.007655 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.007782 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.007800 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.007829 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.007880 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.111816 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.111872 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.111890 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.111914 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.111934 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.215320 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.215368 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.215418 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.215441 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.215459 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.232065 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 10:23:27.535480944 +0000 UTC Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.255650 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.255704 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:30 crc kubenswrapper[4676]: E0124 00:04:30.256008 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:30 crc kubenswrapper[4676]: E0124 00:04:30.256137 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.317999 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.318337 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.318692 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.318978 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.319284 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.422016 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.422079 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.422101 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.422132 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.422154 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.525290 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.525343 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.525360 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.525407 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.525424 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.628206 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.628275 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.628298 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.628323 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.628340 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.731469 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.731809 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.731968 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.732120 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.732250 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.834855 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.834904 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.834921 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.834944 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.834962 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.937455 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.937504 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.937521 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.937544 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:30 crc kubenswrapper[4676]: I0124 00:04:30.937561 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:30Z","lastTransitionTime":"2026-01-24T00:04:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.040227 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.040284 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.040302 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.040321 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.040334 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:31Z","lastTransitionTime":"2026-01-24T00:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.143111 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.143177 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.143195 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.143221 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.143237 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:31Z","lastTransitionTime":"2026-01-24T00:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.283326 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 17:16:20.428641491 +0000 UTC Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.283525 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:31 crc kubenswrapper[4676]: E0124 00:04:31.283719 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.284470 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:31 crc kubenswrapper[4676]: E0124 00:04:31.284646 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.286005 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.286063 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.286081 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.286106 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.286123 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:31Z","lastTransitionTime":"2026-01-24T00:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.389449 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.389535 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.389555 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.389611 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.389632 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:31Z","lastTransitionTime":"2026-01-24T00:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.491910 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.491985 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.492007 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.492035 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.492056 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:31Z","lastTransitionTime":"2026-01-24T00:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.602164 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.602239 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.602261 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.602290 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.602311 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:31Z","lastTransitionTime":"2026-01-24T00:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.704618 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.704668 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.704684 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.704706 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.704723 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:31Z","lastTransitionTime":"2026-01-24T00:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.808125 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.808167 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.808175 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.808189 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.808200 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:31Z","lastTransitionTime":"2026-01-24T00:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.911355 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.911421 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.911434 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.911461 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:31 crc kubenswrapper[4676]: I0124 00:04:31.911480 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:31Z","lastTransitionTime":"2026-01-24T00:04:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.014042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.014102 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.014115 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.014137 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.014155 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.117014 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.117050 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.117059 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.117075 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.117084 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.220588 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.220664 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.220679 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.220702 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.220737 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.255209 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.255236 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:32 crc kubenswrapper[4676]: E0124 00:04:32.255445 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:32 crc kubenswrapper[4676]: E0124 00:04:32.255521 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.284127 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 23:44:22.813873282 +0000 UTC Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.324444 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.324500 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.324510 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.324533 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.324546 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.427529 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.427619 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.427638 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.427663 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.427680 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.531194 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.531330 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.531439 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.531466 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.531484 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.636030 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.636095 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.636109 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.636133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.636149 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.740134 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.740226 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.740246 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.740277 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.740300 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.844066 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.844141 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.844158 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.844188 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.844204 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.947690 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.947746 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.947759 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.947776 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:32 crc kubenswrapper[4676]: I0124 00:04:32.947786 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:32Z","lastTransitionTime":"2026-01-24T00:04:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.051028 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.051100 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.051125 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.051156 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.051179 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.155293 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.155341 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.155356 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.155411 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.155434 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.255607 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.255712 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:33 crc kubenswrapper[4676]: E0124 00:04:33.255788 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:33 crc kubenswrapper[4676]: E0124 00:04:33.255903 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.258877 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.258929 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.258943 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.258970 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.258983 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.285373 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 14:17:57.636103891 +0000 UTC Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.361183 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.361219 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.361229 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.361245 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.361258 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.464530 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.464598 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.464621 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.464655 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.464679 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.566842 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.566894 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.566911 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.566934 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.566952 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.670236 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.670299 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.670321 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.670350 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.670371 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.773109 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.773146 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.773154 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.773184 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.773195 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.875304 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.875364 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.875421 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.875443 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.875456 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.979101 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.979184 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.979201 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.979230 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:33 crc kubenswrapper[4676]: I0124 00:04:33.979248 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:33Z","lastTransitionTime":"2026-01-24T00:04:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.082536 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.082611 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.082623 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.082646 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.082659 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:34Z","lastTransitionTime":"2026-01-24T00:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.185507 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.185548 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.185560 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.185580 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.185592 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:34Z","lastTransitionTime":"2026-01-24T00:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.254851 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.254924 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:34 crc kubenswrapper[4676]: E0124 00:04:34.255444 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:34 crc kubenswrapper[4676]: E0124 00:04:34.255606 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.256005 4676 scope.go:117] "RemoveContainer" containerID="a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76" Jan 24 00:04:34 crc kubenswrapper[4676]: E0124 00:04:34.256252 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.285812 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 21:43:17.99114929 +0000 UTC Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.287427 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.287468 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.287482 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.287500 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.287511 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:34Z","lastTransitionTime":"2026-01-24T00:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.390589 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.390662 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.390682 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.390708 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.390725 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:34Z","lastTransitionTime":"2026-01-24T00:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.493905 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.493947 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.493958 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.493973 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.493983 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:34Z","lastTransitionTime":"2026-01-24T00:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.597142 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.597214 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.597234 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.597263 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.597283 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:34Z","lastTransitionTime":"2026-01-24T00:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.700102 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.700194 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.700213 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.700241 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.700258 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:34Z","lastTransitionTime":"2026-01-24T00:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.803411 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.803502 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.803527 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.803607 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.803641 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:34Z","lastTransitionTime":"2026-01-24T00:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.906926 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.906989 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.907013 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.907042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:34 crc kubenswrapper[4676]: I0124 00:04:34.907065 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:34Z","lastTransitionTime":"2026-01-24T00:04:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.010258 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.010322 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.010333 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.010349 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.010399 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.114172 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.114230 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.114248 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.114273 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.114291 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.217448 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.217503 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.217515 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.217551 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.217564 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.254746 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.255186 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:35 crc kubenswrapper[4676]: E0124 00:04:35.255413 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:35 crc kubenswrapper[4676]: E0124 00:04:35.255543 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.286048 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 05:50:58.616939546 +0000 UTC Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.320354 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.320464 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.320484 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.320874 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.320933 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.424456 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.424591 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.424668 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.424703 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.424944 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.528423 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.528553 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.528578 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.528604 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.528622 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.631756 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.631792 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.631800 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.631814 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.631824 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.734761 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.734820 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.734843 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.734873 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.734896 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.838198 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.838261 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.838281 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.838306 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.838326 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.942081 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.942143 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.942159 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.942185 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:35 crc kubenswrapper[4676]: I0124 00:04:35.942201 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:35Z","lastTransitionTime":"2026-01-24T00:04:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.045582 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.045708 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.045728 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.045752 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.045768 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:36Z","lastTransitionTime":"2026-01-24T00:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.148201 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.148273 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.148283 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.148298 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.148306 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:36Z","lastTransitionTime":"2026-01-24T00:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.251032 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.251536 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.251554 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.251580 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.251597 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:36Z","lastTransitionTime":"2026-01-24T00:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.255105 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:36 crc kubenswrapper[4676]: E0124 00:04:36.255501 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.255878 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:36 crc kubenswrapper[4676]: E0124 00:04:36.256123 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.271388 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.286699 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 08:49:04.533568041 +0000 UTC Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.288104 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.299577 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.312878 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.323785 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.334565 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.342161 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.353607 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.353635 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.353644 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.353656 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.353665 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:36Z","lastTransitionTime":"2026-01-24T00:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.359933 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.372095 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.381983 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.393689 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.403199 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.414961 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.425011 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.433837 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.445683 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.455850 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.456626 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.456670 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.456681 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.456696 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.456706 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:36Z","lastTransitionTime":"2026-01-24T00:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.468633 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:36Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.558546 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.558581 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.558593 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.558610 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.558621 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:36Z","lastTransitionTime":"2026-01-24T00:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.774583 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.774626 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.774641 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.774659 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.774673 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:36Z","lastTransitionTime":"2026-01-24T00:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.877547 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.877602 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.877618 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.877642 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.877659 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:36Z","lastTransitionTime":"2026-01-24T00:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.979694 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.979732 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.979742 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.979757 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:36 crc kubenswrapper[4676]: I0124 00:04:36.979769 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:36Z","lastTransitionTime":"2026-01-24T00:04:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.082368 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.082424 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.082433 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.082449 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.082460 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:37Z","lastTransitionTime":"2026-01-24T00:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.185794 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.185897 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.185920 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.185949 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.185972 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:37Z","lastTransitionTime":"2026-01-24T00:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.255622 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.255755 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:37 crc kubenswrapper[4676]: E0124 00:04:37.255863 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:37 crc kubenswrapper[4676]: E0124 00:04:37.256074 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.287027 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 03:53:09.58734478 +0000 UTC Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.289311 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.289360 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.289405 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.289434 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.289454 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:37Z","lastTransitionTime":"2026-01-24T00:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.392372 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.392922 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.392950 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.392982 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.393010 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:37Z","lastTransitionTime":"2026-01-24T00:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.495702 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.495761 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.495782 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.495811 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.495832 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:37Z","lastTransitionTime":"2026-01-24T00:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.599312 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.599365 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.599409 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.599439 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.599455 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:37Z","lastTransitionTime":"2026-01-24T00:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.703021 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.703480 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.704020 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.704238 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.704475 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:37Z","lastTransitionTime":"2026-01-24T00:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.807994 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.808047 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.808063 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.808087 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.808104 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:37Z","lastTransitionTime":"2026-01-24T00:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.911031 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.911087 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.911108 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.911133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:37 crc kubenswrapper[4676]: I0124 00:04:37.911148 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:37Z","lastTransitionTime":"2026-01-24T00:04:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.013570 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.013631 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.013654 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.013681 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.013704 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.116241 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.116275 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.116286 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.116302 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.116313 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.219681 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.219742 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.219759 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.219782 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.219798 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.255291 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.255291 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:38 crc kubenswrapper[4676]: E0124 00:04:38.255478 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:38 crc kubenswrapper[4676]: E0124 00:04:38.255603 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.288788 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 04:19:27.004281216 +0000 UTC Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.321472 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.321493 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.321501 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.321512 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.321521 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.388586 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.388636 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.388652 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.388674 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.388692 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: E0124 00:04:38.407403 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:38Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.412228 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.412324 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.412341 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.412366 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.412412 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: E0124 00:04:38.430943 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:38Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.435758 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.435821 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.435841 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.435866 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.435883 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: E0124 00:04:38.454687 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:38Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.459712 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.459776 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.459802 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.459834 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.459858 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: E0124 00:04:38.480220 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:38Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.485222 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.485496 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.485699 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.485882 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.486036 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: E0124 00:04:38.505989 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:38Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:38 crc kubenswrapper[4676]: E0124 00:04:38.506739 4676 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.508508 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.508550 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.508562 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.508580 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.508592 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.610729 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.610761 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.610770 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.610782 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.610791 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.714431 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.714491 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.714509 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.714532 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.714548 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.817324 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.817415 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.817435 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.817458 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.817475 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.920617 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.920681 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.920698 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.920721 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:38 crc kubenswrapper[4676]: I0124 00:04:38.920739 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:38Z","lastTransitionTime":"2026-01-24T00:04:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.023926 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.024033 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.024062 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.024103 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.024130 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.127041 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.127339 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.127506 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.127730 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.128026 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.231331 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.231426 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.231441 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.231470 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.231487 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.254994 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.255002 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:39 crc kubenswrapper[4676]: E0124 00:04:39.255227 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:39 crc kubenswrapper[4676]: E0124 00:04:39.255320 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.289133 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 18:54:34.31283418 +0000 UTC Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.334310 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.334432 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.334461 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.334492 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.334511 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.437246 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.437299 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.437312 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.437329 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.437342 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.540726 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.540782 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.540794 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.540818 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.540832 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.643264 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.643316 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.643327 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.643344 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.643356 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.745878 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.745935 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.745955 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.745986 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.746009 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.848874 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.848944 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.848968 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.849026 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.849046 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.951585 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.951639 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.951656 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.951679 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:39 crc kubenswrapper[4676]: I0124 00:04:39.951696 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:39Z","lastTransitionTime":"2026-01-24T00:04:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.054540 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.054630 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.054657 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.054688 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.054710 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:40Z","lastTransitionTime":"2026-01-24T00:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.157211 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.157267 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.157279 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.157300 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.157315 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:40Z","lastTransitionTime":"2026-01-24T00:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.255700 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.255783 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:40 crc kubenswrapper[4676]: E0124 00:04:40.255919 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:40 crc kubenswrapper[4676]: E0124 00:04:40.256068 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.260682 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.260718 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.260726 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.260740 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.260751 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:40Z","lastTransitionTime":"2026-01-24T00:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.289849 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 20:55:07.502831759 +0000 UTC Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.364466 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.364530 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.364542 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.364567 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.364580 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:40Z","lastTransitionTime":"2026-01-24T00:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.583272 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.583308 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.583316 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.583332 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.583341 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:40Z","lastTransitionTime":"2026-01-24T00:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.685258 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.685298 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.685308 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.685325 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.685338 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:40Z","lastTransitionTime":"2026-01-24T00:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.791599 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.791668 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.791690 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.791721 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.791742 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:40Z","lastTransitionTime":"2026-01-24T00:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.893794 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.893829 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.893837 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.893850 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.893859 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:40Z","lastTransitionTime":"2026-01-24T00:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.996053 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.996094 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.996105 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.996122 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:40 crc kubenswrapper[4676]: I0124 00:04:40.996158 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:40Z","lastTransitionTime":"2026-01-24T00:04:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.098337 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.098407 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.098424 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.098443 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.098459 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:41Z","lastTransitionTime":"2026-01-24T00:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.200861 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.200899 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.200908 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.200921 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.200931 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:41Z","lastTransitionTime":"2026-01-24T00:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.254739 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:41 crc kubenswrapper[4676]: E0124 00:04:41.254855 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.255032 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:41 crc kubenswrapper[4676]: E0124 00:04:41.255077 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.290356 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 03:35:56.043579597 +0000 UTC Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.305003 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.305037 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.305048 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.305068 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.305081 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:41Z","lastTransitionTime":"2026-01-24T00:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.407292 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.407329 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.407341 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.407355 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.407365 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:41Z","lastTransitionTime":"2026-01-24T00:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.510623 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.510665 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.510680 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.510701 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.510716 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:41Z","lastTransitionTime":"2026-01-24T00:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.612943 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.612968 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.612975 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.612988 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.612997 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:41Z","lastTransitionTime":"2026-01-24T00:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.715074 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.715114 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.715122 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.715137 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.715147 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:41Z","lastTransitionTime":"2026-01-24T00:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.817161 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.817196 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.817209 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.817224 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.817237 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:41Z","lastTransitionTime":"2026-01-24T00:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.919623 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.919683 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.919700 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.919725 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:41 crc kubenswrapper[4676]: I0124 00:04:41.919747 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:41Z","lastTransitionTime":"2026-01-24T00:04:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.022133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.022170 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.022182 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.022201 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.022213 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.124966 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.125013 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.125024 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.125043 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.125054 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.227854 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.227892 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.227900 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.227915 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.227924 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.255593 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.255659 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:42 crc kubenswrapper[4676]: E0124 00:04:42.255698 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:42 crc kubenswrapper[4676]: E0124 00:04:42.255856 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.291204 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 07:16:12.407153719 +0000 UTC Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.330340 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.330400 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.330410 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.330429 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.330440 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.331136 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:42 crc kubenswrapper[4676]: E0124 00:04:42.331338 4676 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:42 crc kubenswrapper[4676]: E0124 00:04:42.331463 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs podName:18335446-e572-4741-ad9e-e7aadee7550b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:14.331435898 +0000 UTC m=+98.361406939 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs") pod "network-metrics-daemon-r4q22" (UID: "18335446-e572-4741-ad9e-e7aadee7550b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.432518 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.432562 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.432571 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.432590 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.432602 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.534148 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.534182 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.534190 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.534203 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.534212 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.636462 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.636497 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.636506 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.636522 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.636531 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.738872 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.738900 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.738909 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.738923 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.738932 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.793067 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/0.log" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.793115 4676 generic.go:334] "Generic (PLEG): container finished" podID="b88e9d2e-35da-45a8-ac7e-22afd660ff9f" containerID="db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c" exitCode=1 Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.793148 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x57xf" event={"ID":"b88e9d2e-35da-45a8-ac7e-22afd660ff9f","Type":"ContainerDied","Data":"db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.793481 4676 scope.go:117] "RemoveContainer" containerID="db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.804506 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.819704 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.829720 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.840789 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"2026-01-24T00:03:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf\\\\n2026-01-24T00:03:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf to /host/opt/cni/bin/\\\\n2026-01-24T00:03:57Z [verbose] multus-daemon started\\\\n2026-01-24T00:03:57Z [verbose] Readiness Indicator file check\\\\n2026-01-24T00:04:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.841094 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.841122 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.841132 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.841148 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.841158 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.854851 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.873369 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.884532 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.899388 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.911823 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.924238 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.942863 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.944098 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.944140 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.944339 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.944360 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.944411 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:42Z","lastTransitionTime":"2026-01-24T00:04:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.953662 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.964838 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.978858 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:42 crc kubenswrapper[4676]: I0124 00:04:42.989524 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:42Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.011803 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.025240 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.040079 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.047467 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.047501 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.047511 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.047526 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.047536 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.150691 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.150728 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.150737 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.150752 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.150761 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.253427 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.253489 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.253506 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.253530 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.253547 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.255555 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.255597 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:43 crc kubenswrapper[4676]: E0124 00:04:43.255645 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:43 crc kubenswrapper[4676]: E0124 00:04:43.255719 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.291761 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 01:06:34.02742757 +0000 UTC Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.355719 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.355760 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.355774 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.355792 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.355804 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.457947 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.458007 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.458018 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.458033 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.458044 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.559836 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.559898 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.559915 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.559937 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.559954 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.662233 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.662275 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.662284 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.662300 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.662310 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.764570 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.764610 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.764621 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.764640 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.764650 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.799161 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/0.log" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.799466 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x57xf" event={"ID":"b88e9d2e-35da-45a8-ac7e-22afd660ff9f","Type":"ContainerStarted","Data":"cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.813176 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.823810 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.848669 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.862490 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.867090 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.867149 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.867166 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.867189 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.867206 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.876874 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.891087 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.905694 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.918726 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.944324 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.957233 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.970573 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.970611 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.970625 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.970643 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.970656 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:43Z","lastTransitionTime":"2026-01-24T00:04:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.971663 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:43 crc kubenswrapper[4676]: I0124 00:04:43.984389 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.000161 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:43Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.018186 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:44Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.037243 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"2026-01-24T00:03:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf\\\\n2026-01-24T00:03:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf to /host/opt/cni/bin/\\\\n2026-01-24T00:03:57Z [verbose] multus-daemon started\\\\n2026-01-24T00:03:57Z [verbose] Readiness Indicator file check\\\\n2026-01-24T00:04:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:44Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.053600 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:44Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.070200 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:44Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.073073 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.073125 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.073144 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.073170 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.073189 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:44Z","lastTransitionTime":"2026-01-24T00:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.085543 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:44Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.176296 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.176642 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.176810 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.176960 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.177108 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:44Z","lastTransitionTime":"2026-01-24T00:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.254717 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:44 crc kubenswrapper[4676]: E0124 00:04:44.254901 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.255296 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:44 crc kubenswrapper[4676]: E0124 00:04:44.255858 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.292282 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 01:51:55.260539834 +0000 UTC Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.312921 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.312971 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.312984 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.313003 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.313017 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:44Z","lastTransitionTime":"2026-01-24T00:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.415626 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.415836 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.415895 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.415969 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.416029 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:44Z","lastTransitionTime":"2026-01-24T00:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.518367 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.518431 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.518443 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.518462 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.518474 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:44Z","lastTransitionTime":"2026-01-24T00:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.620684 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.620719 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.620730 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.620747 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.620758 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:44Z","lastTransitionTime":"2026-01-24T00:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.723398 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.723428 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.723436 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.723453 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.723460 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:44Z","lastTransitionTime":"2026-01-24T00:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.825103 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.825423 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.825512 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.825621 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.825703 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:44Z","lastTransitionTime":"2026-01-24T00:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.928478 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.928569 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.928582 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.928603 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:44 crc kubenswrapper[4676]: I0124 00:04:44.928617 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:44Z","lastTransitionTime":"2026-01-24T00:04:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.031441 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.031503 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.031522 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.031543 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.031557 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.134213 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.134248 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.134258 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.134273 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.134282 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.236501 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.236529 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.236539 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.236550 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.236558 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.255264 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.255313 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:45 crc kubenswrapper[4676]: E0124 00:04:45.255443 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:45 crc kubenswrapper[4676]: E0124 00:04:45.255588 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.293785 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 01:23:31.889915031 +0000 UTC Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.338350 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.338376 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.338384 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.338397 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.338405 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.440678 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.440707 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.440718 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.440733 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.440743 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.543055 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.543097 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.543114 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.543137 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.543156 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.645807 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.645840 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.645852 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.645870 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.645884 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.748035 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.748063 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.748071 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.748086 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.748096 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.850815 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.850875 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.850892 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.851143 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.851164 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.954534 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.954677 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.954700 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.954729 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:45 crc kubenswrapper[4676]: I0124 00:04:45.954752 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:45Z","lastTransitionTime":"2026-01-24T00:04:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.057777 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.057816 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.057825 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.057840 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.057851 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.159942 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.159985 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.159997 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.160018 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.160034 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.255925 4676 scope.go:117] "RemoveContainer" containerID="a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.256356 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:46 crc kubenswrapper[4676]: E0124 00:04:46.256438 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.256610 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:46 crc kubenswrapper[4676]: E0124 00:04:46.256689 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.265982 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.266098 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.266172 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.266250 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.266316 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.273824 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.294782 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 23:45:54.708886724 +0000 UTC Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.297022 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.316667 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.328095 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.356109 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.369131 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.369164 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.369174 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.369187 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.369197 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.371008 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.382150 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.393851 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.407345 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.424043 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.438161 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"2026-01-24T00:03:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf\\\\n2026-01-24T00:03:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf to /host/opt/cni/bin/\\\\n2026-01-24T00:03:57Z [verbose] multus-daemon started\\\\n2026-01-24T00:03:57Z [verbose] Readiness Indicator file check\\\\n2026-01-24T00:04:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.451402 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.463880 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.470947 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.471123 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.471234 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.471348 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.471445 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.478309 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.494583 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.508500 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.528981 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.540284 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.573360 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.573572 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.573658 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.573720 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.573775 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.676080 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.676327 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.676439 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.676524 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.676608 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.778940 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.778983 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.778993 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.779009 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.779018 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.810896 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/2.log" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.815506 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.816251 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.827003 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.843943 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.859192 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.870291 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.884742 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.884791 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.884803 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.884821 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.884839 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.888796 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.905030 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.921750 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.936364 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.951224 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.961535 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.972543 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"2026-01-24T00:03:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf\\\\n2026-01-24T00:03:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf to /host/opt/cni/bin/\\\\n2026-01-24T00:03:57Z [verbose] multus-daemon started\\\\n2026-01-24T00:03:57Z [verbose] Readiness Indicator file check\\\\n2026-01-24T00:04:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.984315 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.986588 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.986620 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.986629 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.986643 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.986653 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:46Z","lastTransitionTime":"2026-01-24T00:04:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:46 crc kubenswrapper[4676]: I0124 00:04:46.997712 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:46Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.008174 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.019135 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.029319 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.046683 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.058331 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.088604 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.088630 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.088638 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.088650 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.088659 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:47Z","lastTransitionTime":"2026-01-24T00:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.190993 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.191047 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.191065 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.191088 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.191104 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:47Z","lastTransitionTime":"2026-01-24T00:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.254816 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.254816 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:47 crc kubenswrapper[4676]: E0124 00:04:47.254939 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:47 crc kubenswrapper[4676]: E0124 00:04:47.255015 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.293050 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.293084 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.293092 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.293106 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.293115 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:47Z","lastTransitionTime":"2026-01-24T00:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.295428 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 14:55:12.077624838 +0000 UTC Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.394627 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.394672 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.394683 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.394700 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.394711 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:47Z","lastTransitionTime":"2026-01-24T00:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.496798 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.496833 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.496841 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.496854 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.496864 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:47Z","lastTransitionTime":"2026-01-24T00:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.599530 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.599571 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.599579 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.599596 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.599606 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:47Z","lastTransitionTime":"2026-01-24T00:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.701066 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.701106 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.701114 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.701127 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.701136 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:47Z","lastTransitionTime":"2026-01-24T00:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.803655 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.803720 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.803738 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.803771 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.803802 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:47Z","lastTransitionTime":"2026-01-24T00:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.819877 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/3.log" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.820506 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/2.log" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.823858 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9" exitCode=1 Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.823896 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.823959 4676 scope.go:117] "RemoveContainer" containerID="a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.824563 4676 scope.go:117] "RemoveContainer" containerID="dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9" Jan 24 00:04:47 crc kubenswrapper[4676]: E0124 00:04:47.824734 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.840423 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.852466 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.877004 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59fa26ff296d45c03abf7bef737940a6dadc3ec78d1b4e3b43b52a803530a76\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:21Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:20.427464 6215 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0124 00:04:20.427560 6215 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0124 00:04:20.427590 6215 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0124 00:04:20.427626 6215 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0124 00:04:20.427633 6215 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0124 00:04:20.427665 6215 handler.go:208] Removed *v1.Node event handler 7\\\\nI0124 00:04:20.427671 6215 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0124 00:04:20.427693 6215 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0124 00:04:20.427707 6215 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0124 00:04:20.427719 6215 handler.go:208] Removed *v1.Node event handler 2\\\\nI0124 00:04:20.427714 6215 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0124 00:04:20.427729 6215 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0124 00:04:20.427751 6215 factory.go:656] Stopping watch factory\\\\nI0124 00:04:20.427768 6215 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:20.427795 6215 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0124 00:04:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:47Z\\\",\\\"message\\\":\\\" reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.034512 6579 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035026 6579 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035147 6579 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035214 6579 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0124 00:04:47.035267 6579 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0124 00:04:47.040579 6579 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0124 00:04:47.040623 6579 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0124 00:04:47.040714 6579 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:47.040756 6579 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0124 00:04:47.040876 6579 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.891329 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.906077 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.906117 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.906127 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.906143 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.906152 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:47Z","lastTransitionTime":"2026-01-24T00:04:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.919167 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.931990 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.942798 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.953253 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.965084 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.976633 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.987368 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"2026-01-24T00:03:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf\\\\n2026-01-24T00:03:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf to /host/opt/cni/bin/\\\\n2026-01-24T00:03:57Z [verbose] multus-daemon started\\\\n2026-01-24T00:03:57Z [verbose] Readiness Indicator file check\\\\n2026-01-24T00:04:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:47 crc kubenswrapper[4676]: I0124 00:04:47.997875 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:47Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.008091 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.008133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.008144 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.008160 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.008171 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.014151 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.030880 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.041274 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.054350 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.069745 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.081509 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.110288 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.110315 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.110323 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.110336 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.110344 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.213178 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.213245 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.213268 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.213295 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.213317 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.254718 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.254718 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:48 crc kubenswrapper[4676]: E0124 00:04:48.254956 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:48 crc kubenswrapper[4676]: E0124 00:04:48.255076 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.295976 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 20:11:38.740752422 +0000 UTC Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.315669 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.315714 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.315728 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.315747 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.315761 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.417313 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.417392 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.417405 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.417420 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.417431 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.520050 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.520100 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.520112 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.520130 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.520145 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.622482 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.622536 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.622551 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.622572 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.622587 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.666220 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.666274 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.666284 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.666303 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.666317 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: E0124 00:04:48.677630 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.680935 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.680955 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.680963 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.680977 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.680989 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: E0124 00:04:48.694367 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.697935 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.697958 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.697968 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.697982 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.697992 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: E0124 00:04:48.708645 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.712337 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.712382 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.712392 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.712424 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.712435 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: E0124 00:04:48.724429 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.728305 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.728338 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.728348 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.728364 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.728376 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: E0124 00:04:48.740310 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: E0124 00:04:48.740445 4676 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.741825 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.741850 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.741861 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.741872 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.741881 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.828064 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/3.log" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.832003 4676 scope.go:117] "RemoveContainer" containerID="dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9" Jan 24 00:04:48 crc kubenswrapper[4676]: E0124 00:04:48.832266 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.843747 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.843769 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.843778 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.843792 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.843801 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.847045 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.864537 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.876835 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.890218 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"2026-01-24T00:03:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf\\\\n2026-01-24T00:03:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf to /host/opt/cni/bin/\\\\n2026-01-24T00:03:57Z [verbose] multus-daemon started\\\\n2026-01-24T00:03:57Z [verbose] Readiness Indicator file check\\\\n2026-01-24T00:04:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.906154 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.925073 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.939845 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.945973 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.946015 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.946042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.946060 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.946070 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:48Z","lastTransitionTime":"2026-01-24T00:04:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.955136 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.971221 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:48 crc kubenswrapper[4676]: I0124 00:04:48.983181 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:48Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.004241 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:47Z\\\",\\\"message\\\":\\\" reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.034512 6579 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035026 6579 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035147 6579 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035214 6579 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0124 00:04:47.035267 6579 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0124 00:04:47.040579 6579 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0124 00:04:47.040623 6579 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0124 00:04:47.040714 6579 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:47.040756 6579 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0124 00:04:47.040876 6579 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:49Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.016541 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:49Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.032178 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:49Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.053909 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.053945 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.053963 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.053980 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.053992 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.054200 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:49Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.071575 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:49Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.101156 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:49Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.115733 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:49Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.134865 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:49Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.156879 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.156920 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.156929 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.156945 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.156955 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.255574 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.255619 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:49 crc kubenswrapper[4676]: E0124 00:04:49.255714 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:49 crc kubenswrapper[4676]: E0124 00:04:49.255871 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.259039 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.259071 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.259082 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.259096 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.259107 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.296147 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 07:42:39.89309252 +0000 UTC Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.372282 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.372315 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.372326 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.372344 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.372356 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.475022 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.475100 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.475124 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.475157 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.475181 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.578031 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.578066 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.578075 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.578091 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.578100 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.680887 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.680936 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.680954 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.680977 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.680993 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.783848 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.783904 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.783920 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.783945 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.783967 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.886952 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.887019 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.887036 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.887064 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.887084 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.990828 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.990888 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.990906 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.990931 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:49 crc kubenswrapper[4676]: I0124 00:04:49.990948 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:49Z","lastTransitionTime":"2026-01-24T00:04:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.093293 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.093336 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.093348 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.093366 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.093402 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:50Z","lastTransitionTime":"2026-01-24T00:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.195874 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.195934 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.195951 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.195979 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.195996 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:50Z","lastTransitionTime":"2026-01-24T00:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.254824 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.254835 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:50 crc kubenswrapper[4676]: E0124 00:04:50.254938 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:50 crc kubenswrapper[4676]: E0124 00:04:50.255075 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.296964 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 15:58:47.965914184 +0000 UTC Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.299670 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.299727 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.299744 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.299768 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.299785 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:50Z","lastTransitionTime":"2026-01-24T00:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.402178 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.402224 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.402235 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.402250 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.402260 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:50Z","lastTransitionTime":"2026-01-24T00:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.505169 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.505207 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.505217 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.505233 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.505244 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:50Z","lastTransitionTime":"2026-01-24T00:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.608811 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.608869 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.608890 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.608918 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.608939 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:50Z","lastTransitionTime":"2026-01-24T00:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.712613 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.712683 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.712704 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.712736 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.712761 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:50Z","lastTransitionTime":"2026-01-24T00:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.820450 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.820504 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.820523 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.820549 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.820567 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:50Z","lastTransitionTime":"2026-01-24T00:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.922825 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.922866 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.922877 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.922892 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:50 crc kubenswrapper[4676]: I0124 00:04:50.922903 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:50Z","lastTransitionTime":"2026-01-24T00:04:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.025537 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.025595 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.025612 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.025636 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.025653 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.127701 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.127746 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.127757 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.127774 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.127788 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.230750 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.230815 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.230833 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.230859 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.230876 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.255459 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:51 crc kubenswrapper[4676]: E0124 00:04:51.255623 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.255468 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:51 crc kubenswrapper[4676]: E0124 00:04:51.255878 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.297609 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 09:49:26.616841929 +0000 UTC Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.334493 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.334564 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.334586 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.334617 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.334636 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.437064 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.437493 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.437519 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.437542 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.437559 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.540529 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.540565 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.540576 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.540590 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.540619 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.643599 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.643938 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.644092 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.644245 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.644425 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.747876 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.747932 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.747943 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.747961 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.747973 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.850706 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.851091 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.851171 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.851266 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.851356 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.954533 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.954917 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.954996 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.955068 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:51 crc kubenswrapper[4676]: I0124 00:04:51.955142 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:51Z","lastTransitionTime":"2026-01-24T00:04:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.058792 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.058857 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.058877 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.058905 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.058924 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.162941 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.163324 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.163412 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.163510 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.163613 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.255616 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.255729 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:52 crc kubenswrapper[4676]: E0124 00:04:52.256495 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:52 crc kubenswrapper[4676]: E0124 00:04:52.256707 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.265519 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.265762 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.265901 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.266050 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.266338 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.297827 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 00:46:17.728723874 +0000 UTC Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.369756 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.369796 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.369807 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.369824 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.369835 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.473020 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.473074 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.473086 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.473106 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.473119 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.577657 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.577726 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.577746 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.577772 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.577791 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.680504 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.681004 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.681201 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.681334 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.681513 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.784684 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.784742 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.784759 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.784786 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.784804 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.887693 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.887762 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.887779 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.887804 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.887822 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.990557 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.991021 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.991334 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.991669 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:52 crc kubenswrapper[4676]: I0124 00:04:52.991811 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:52Z","lastTransitionTime":"2026-01-24T00:04:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.094630 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.094901 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.095048 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.095192 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.095308 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:53Z","lastTransitionTime":"2026-01-24T00:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.198133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.198442 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.198588 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.198733 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.198873 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:53Z","lastTransitionTime":"2026-01-24T00:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.255729 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:53 crc kubenswrapper[4676]: E0124 00:04:53.255903 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.255749 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:53 crc kubenswrapper[4676]: E0124 00:04:53.256343 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.298819 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 08:02:20.73729298 +0000 UTC Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.301625 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.301820 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.302217 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.302457 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.302890 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:53Z","lastTransitionTime":"2026-01-24T00:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.405999 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.406249 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.406475 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.406658 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.406799 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:53Z","lastTransitionTime":"2026-01-24T00:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.508752 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.509430 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.509906 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.510407 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.510595 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:53Z","lastTransitionTime":"2026-01-24T00:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.613309 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.613394 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.613414 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.613445 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.613469 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:53Z","lastTransitionTime":"2026-01-24T00:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.715952 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.716002 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.716011 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.716025 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.716034 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:53Z","lastTransitionTime":"2026-01-24T00:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.819066 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.819124 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.819141 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.819165 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.819182 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:53Z","lastTransitionTime":"2026-01-24T00:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.922578 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.922648 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.922669 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.922698 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:53 crc kubenswrapper[4676]: I0124 00:04:53.923149 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:53Z","lastTransitionTime":"2026-01-24T00:04:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.027087 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.027135 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.027146 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.027165 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.027176 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.130267 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.130399 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.130420 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.130445 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.130463 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.233237 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.233326 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.233343 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.233368 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.233417 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.255083 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:54 crc kubenswrapper[4676]: E0124 00:04:54.255330 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.255086 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:54 crc kubenswrapper[4676]: E0124 00:04:54.255663 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.299284 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 10:08:47.956276261 +0000 UTC Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.335827 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.335901 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.335919 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.335945 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.335963 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.439634 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.439682 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.439980 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.440006 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.440023 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.543333 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.543461 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.543488 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.543517 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.543538 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.646465 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.646531 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.646549 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.646573 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.646590 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.749742 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.749800 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.749816 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.749839 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.749856 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.852336 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.852433 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.852454 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.852479 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.852497 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.957324 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.957444 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.957470 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.957500 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:54 crc kubenswrapper[4676]: I0124 00:04:54.957522 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:54Z","lastTransitionTime":"2026-01-24T00:04:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.061029 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.061082 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.061098 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.061119 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.061135 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.164095 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.164141 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.164159 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.164182 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.164199 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.254959 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.254974 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:55 crc kubenswrapper[4676]: E0124 00:04:55.255179 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:55 crc kubenswrapper[4676]: E0124 00:04:55.255259 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.267299 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.267368 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.267437 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.267477 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.267501 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.300014 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 11:38:12.187481517 +0000 UTC Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.370242 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.370284 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.370297 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.370314 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.370325 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.473279 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.473338 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.473356 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.473378 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.473422 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.575770 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.575819 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.575831 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.575846 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.575856 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.677740 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.677779 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.677789 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.677805 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.677817 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.780127 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.780172 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.780190 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.780213 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.780231 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.883170 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.883301 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.883328 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.883356 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.883383 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.985857 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.985900 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.985911 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.985929 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:55 crc kubenswrapper[4676]: I0124 00:04:55.985942 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:55Z","lastTransitionTime":"2026-01-24T00:04:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.089454 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.089502 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.089519 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.089541 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.089557 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:56Z","lastTransitionTime":"2026-01-24T00:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.192454 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.192514 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.192594 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.192627 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.192649 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:56Z","lastTransitionTime":"2026-01-24T00:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.259191 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:56 crc kubenswrapper[4676]: E0124 00:04:56.260077 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.259696 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:56 crc kubenswrapper[4676]: E0124 00:04:56.260450 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.281984 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.298133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.298398 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.298503 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.298588 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.298675 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:56Z","lastTransitionTime":"2026-01-24T00:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.300492 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 09:26:26.330485113 +0000 UTC Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.303758 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.328810 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"2026-01-24T00:03:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf\\\\n2026-01-24T00:03:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf to /host/opt/cni/bin/\\\\n2026-01-24T00:03:57Z [verbose] multus-daemon started\\\\n2026-01-24T00:03:57Z [verbose] Readiness Indicator file check\\\\n2026-01-24T00:04:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.352919 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.373800 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.389990 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.401634 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.401667 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.401679 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.401693 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.401797 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:56Z","lastTransitionTime":"2026-01-24T00:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.401759 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.410653 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.420077 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.428653 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.449135 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:47Z\\\",\\\"message\\\":\\\" reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.034512 6579 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035026 6579 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035147 6579 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035214 6579 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0124 00:04:47.035267 6579 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0124 00:04:47.040579 6579 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0124 00:04:47.040623 6579 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0124 00:04:47.040714 6579 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:47.040756 6579 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0124 00:04:47.040876 6579 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.460071 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.475952 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.487907 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.504313 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.504395 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.504408 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.504424 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.504435 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:56Z","lastTransitionTime":"2026-01-24T00:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.507602 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.522664 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.534241 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.548920 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:56Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.608345 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.608464 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.608493 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.608525 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.608550 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:56Z","lastTransitionTime":"2026-01-24T00:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.712159 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.712192 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.712205 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.712221 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.712232 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:56Z","lastTransitionTime":"2026-01-24T00:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.840045 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.840083 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.840095 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.840111 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.840122 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:56Z","lastTransitionTime":"2026-01-24T00:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.942258 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.942295 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.942305 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.942322 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:56 crc kubenswrapper[4676]: I0124 00:04:56.942334 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:56Z","lastTransitionTime":"2026-01-24T00:04:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.045373 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.045452 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.045469 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.045492 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.045508 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.148569 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.148635 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.148656 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.148685 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.148739 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.253054 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.253086 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.253095 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.253109 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.253118 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.255509 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.255583 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:57 crc kubenswrapper[4676]: E0124 00:04:57.255622 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:57 crc kubenswrapper[4676]: E0124 00:04:57.255731 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.301217 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 04:09:25.171810413 +0000 UTC Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.356076 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.356435 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.356660 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.356854 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.357044 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.460458 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.460500 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.460517 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.460539 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.460555 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.563133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.563206 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.563229 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.563257 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.563278 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.666220 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.666284 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.666302 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.666330 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.666346 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.768752 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.768808 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.768830 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.768858 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.768881 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.871527 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.871587 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.871609 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.871636 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.871658 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.974497 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.974558 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.974576 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.974600 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:57 crc kubenswrapper[4676]: I0124 00:04:57.974619 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:57Z","lastTransitionTime":"2026-01-24T00:04:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.077164 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.077801 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.078005 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.078107 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.078197 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.180219 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.180259 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.180269 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.180285 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.180295 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.254768 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:58 crc kubenswrapper[4676]: E0124 00:04:58.254983 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.254782 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:04:58 crc kubenswrapper[4676]: E0124 00:04:58.255365 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.283069 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.283119 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.283136 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.283161 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.283177 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.301589 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 07:18:23.640100867 +0000 UTC Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.386577 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.386602 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.386609 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.386623 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.386631 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.489819 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.490222 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.490445 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.490674 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.490874 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.593653 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.594042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.594227 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.594485 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.594688 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.697631 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.698068 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.698223 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.698372 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.698564 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.802028 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.802101 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.802131 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.802162 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.802184 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.866266 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.866322 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.866339 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.866362 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.866411 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: E0124 00:04:58.887830 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.893355 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.893437 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.893455 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.893479 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.893497 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: E0124 00:04:58.916271 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.921669 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.921720 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.921738 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.921762 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.921781 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: E0124 00:04:58.942416 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.954615 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.954678 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.954697 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.954723 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.954741 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:58 crc kubenswrapper[4676]: E0124 00:04:58.976479 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.981749 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.981801 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.981819 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.981848 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:58 crc kubenswrapper[4676]: I0124 00:04:58.981867 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:58Z","lastTransitionTime":"2026-01-24T00:04:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.001776 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:04:58Z is after 2025-08-24T17:21:41Z" Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.002107 4676 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.004435 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.004495 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.004512 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.004537 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.004553 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.107218 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.107268 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.107293 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.107323 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.107345 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.210819 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.210885 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.210908 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.210934 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.210954 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.221348 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.221523 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.221494725 +0000 UTC m=+147.251465756 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.254666 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.254816 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.255054 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.255149 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.271050 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.302625 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 16:42:36.806324763 +0000 UTC Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.314859 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.314914 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.314938 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.314967 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.314989 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.322825 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.322889 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.323001 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323056 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323088 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323107 4676 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323168 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.32314604 +0000 UTC m=+147.353117081 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323170 4676 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323218 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.323206302 +0000 UTC m=+147.353177343 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.323052 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323279 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323308 4676 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323330 4676 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323437 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.323412968 +0000 UTC m=+147.353384000 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323516 4676 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:04:59 crc kubenswrapper[4676]: E0124 00:04:59.323578 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.323558903 +0000 UTC m=+147.353529944 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.418243 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.418298 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.418317 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.418344 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.418361 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.522136 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.522193 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.522209 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.522233 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.522250 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.624989 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.625080 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.625098 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.625120 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.625138 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.727646 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.727709 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.727728 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.727754 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.727777 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.831111 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.831146 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.831157 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.831174 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.831187 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.934519 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.934801 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.934825 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.934857 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:04:59 crc kubenswrapper[4676]: I0124 00:04:59.934879 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:04:59Z","lastTransitionTime":"2026-01-24T00:04:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.038230 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.038288 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.038305 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.038327 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.038342 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.142724 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.142775 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.142792 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.142821 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.142838 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.245822 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.245867 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.245881 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.245901 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.245914 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.255672 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.255720 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:00 crc kubenswrapper[4676]: E0124 00:05:00.256112 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:00 crc kubenswrapper[4676]: E0124 00:05:00.256319 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.304429 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 13:42:23.399296239 +0000 UTC Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.349416 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.349503 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.349525 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.349553 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.349574 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.453194 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.453255 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.453273 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.453296 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.453315 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.555612 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.555708 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.555733 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.555761 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.555784 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.659018 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.659072 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.659090 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.659117 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.659134 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.761951 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.762027 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.762045 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.762070 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.762086 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.864847 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.864915 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.864928 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.864943 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.864954 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.967286 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.967362 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.967414 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.967444 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:00 crc kubenswrapper[4676]: I0124 00:05:00.967462 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:00Z","lastTransitionTime":"2026-01-24T00:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.070118 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.070150 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.070158 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.070184 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.070193 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.173446 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.173505 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.173524 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.173548 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.173565 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.255863 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:01 crc kubenswrapper[4676]: E0124 00:05:01.256120 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.257213 4676 scope.go:117] "RemoveContainer" containerID="dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9" Jan 24 00:05:01 crc kubenswrapper[4676]: E0124 00:05:01.257555 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.257631 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:01 crc kubenswrapper[4676]: E0124 00:05:01.257744 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.277719 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.277768 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.277786 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.277810 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.277828 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.305262 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 10:28:02.044116927 +0000 UTC Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.380676 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.380745 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.380763 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.380787 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.380805 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.483273 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.483336 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.483353 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.483404 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.483424 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.586106 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.586221 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.586245 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.586276 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.586295 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.689464 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.689519 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.689536 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.689558 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.689577 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.792220 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.792273 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.792290 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.792313 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.792329 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.894570 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.894615 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.894631 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.894652 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.894670 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.997955 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.998018 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.998040 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.998068 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:01 crc kubenswrapper[4676]: I0124 00:05:01.998091 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:01Z","lastTransitionTime":"2026-01-24T00:05:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.101096 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.101166 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.101191 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.101222 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.101239 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:02Z","lastTransitionTime":"2026-01-24T00:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.204084 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.204170 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.204195 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.204223 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.204246 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:02Z","lastTransitionTime":"2026-01-24T00:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.255467 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.255508 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:02 crc kubenswrapper[4676]: E0124 00:05:02.255840 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:02 crc kubenswrapper[4676]: E0124 00:05:02.255952 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.305407 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 00:50:39.664224188 +0000 UTC Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.306983 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.307034 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.307053 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.307074 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.307089 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:02Z","lastTransitionTime":"2026-01-24T00:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.410317 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.410406 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.410432 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.410462 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.410485 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:02Z","lastTransitionTime":"2026-01-24T00:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.513702 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.513772 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.513786 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.513810 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.513825 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:02Z","lastTransitionTime":"2026-01-24T00:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.616975 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.617022 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.617037 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.617059 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.617075 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:02Z","lastTransitionTime":"2026-01-24T00:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.719594 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.719648 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.719678 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.719704 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.719720 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:02Z","lastTransitionTime":"2026-01-24T00:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.823018 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.823185 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.823212 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.823243 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.823264 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:02Z","lastTransitionTime":"2026-01-24T00:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.925757 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.925817 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.925835 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.925866 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:02 crc kubenswrapper[4676]: I0124 00:05:02.925884 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:02Z","lastTransitionTime":"2026-01-24T00:05:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.030121 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.030197 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.030217 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.030251 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.030273 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.132986 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.133051 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.133070 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.133094 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.133112 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.236202 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.236338 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.236367 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.236439 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.236466 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.255319 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.255436 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:03 crc kubenswrapper[4676]: E0124 00:05:03.255527 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:03 crc kubenswrapper[4676]: E0124 00:05:03.255643 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.306078 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 03:39:25.845784725 +0000 UTC Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.340368 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.340455 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.340494 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.340526 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.340548 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.443038 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.443108 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.443137 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.443167 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.443191 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.545629 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.545759 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.545784 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.545814 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.545840 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.648577 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.648628 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.648645 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.648666 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.648683 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.751561 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.751592 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.751601 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.751616 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.751627 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.854242 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.854408 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.854440 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.854500 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.854519 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.957636 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.957685 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.957697 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.957713 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:03 crc kubenswrapper[4676]: I0124 00:05:03.957724 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:03Z","lastTransitionTime":"2026-01-24T00:05:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.060552 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.060608 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.060621 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.060641 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.060654 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.163615 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.163701 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.163723 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.163753 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.163778 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.255909 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.256162 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:04 crc kubenswrapper[4676]: E0124 00:05:04.256403 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:04 crc kubenswrapper[4676]: E0124 00:05:04.256563 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.266517 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.266667 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.266692 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.266718 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.266736 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.307149 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 08:21:48.503218661 +0000 UTC Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.369663 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.369764 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.369798 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.369839 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.369866 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.472911 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.472958 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.472975 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.472998 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.473017 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.576988 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.577050 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.577062 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.577085 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.577100 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.679203 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.679247 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.679257 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.679277 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.679287 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.782905 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.783006 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.783074 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.783107 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.783127 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.885789 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.885854 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.885871 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.885898 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.885916 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.988063 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.988101 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.988113 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.988128 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:04 crc kubenswrapper[4676]: I0124 00:05:04.988138 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:04Z","lastTransitionTime":"2026-01-24T00:05:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.091791 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.091859 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.091883 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.091912 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.091931 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:05Z","lastTransitionTime":"2026-01-24T00:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.194177 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.194257 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.194281 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.194314 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.194337 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:05Z","lastTransitionTime":"2026-01-24T00:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.254969 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:05 crc kubenswrapper[4676]: E0124 00:05:05.255143 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.255203 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:05 crc kubenswrapper[4676]: E0124 00:05:05.255711 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.296646 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.296705 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.296723 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.296749 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.296768 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:05Z","lastTransitionTime":"2026-01-24T00:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.308112 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 16:12:05.688953167 +0000 UTC Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.399488 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.399556 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.399577 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.399603 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.399622 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:05Z","lastTransitionTime":"2026-01-24T00:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.502197 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.502248 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.502265 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.502288 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.502309 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:05Z","lastTransitionTime":"2026-01-24T00:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.604784 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.604829 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.604846 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.604868 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.604885 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:05Z","lastTransitionTime":"2026-01-24T00:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.707530 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.707626 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.707645 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.707668 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.707685 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:05Z","lastTransitionTime":"2026-01-24T00:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.810043 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.810094 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.810113 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.810135 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.810152 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:05Z","lastTransitionTime":"2026-01-24T00:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.913595 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.913646 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.913662 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.913686 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:05 crc kubenswrapper[4676]: I0124 00:05:05.913705 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:05Z","lastTransitionTime":"2026-01-24T00:05:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.016593 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.016653 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.016671 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.016695 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.016715 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.120487 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.120552 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.120571 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.120599 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.120616 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.223542 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.223629 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.223648 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.223671 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.223687 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.255279 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:06 crc kubenswrapper[4676]: E0124 00:05:06.255468 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.255709 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:06 crc kubenswrapper[4676]: E0124 00:05:06.255826 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.279735 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24f0dc26-0857-430f-aebd-073fcfcc1c0a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:47Z\\\",\\\"message\\\":\\\" reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.034512 6579 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035026 6579 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035147 6579 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0124 00:04:47.035214 6579 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI0124 00:04:47.035267 6579 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0124 00:04:47.040579 6579 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0124 00:04:47.040623 6579 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0124 00:04:47.040714 6579 ovnkube.go:599] Stopped ovnkube\\\\nI0124 00:04:47.040756 6579 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0124 00:04:47.040876 6579 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4frqp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-ld569\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.298080 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"40151406-46c7-4668-8b2b-db0585847be9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d935e1a0b95e7b7bdbd9c5299727d3f056f62ab78b0062468dac8a66196e023\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7097d31bd127d1e68680dfec923eecc06e9a43f0cf00153752e237b0c013d39d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sj4f8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7m8ts\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.309263 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 18:58:54.06456004 +0000 UTC Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.318164 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.326733 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.326775 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.326791 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.326814 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.326831 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.341094 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"79ad333b-cf18-4ba3-b9d4-2f89c7c44354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7b6909340c11df3cad3b601def65f1a29ff042dd39375eb985c8c8e29442cc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0fd71005d4fd7fe142e31233e3e9aef36b280e5f7531c46df616bba8ef261cbe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dec4e1df33b002745ef1956312b8f1e0ef6b041fa7c8641cc718d26ffc545ac9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6588cf5ad649e704406a3c7d0d036d9913a13d0bdd14f726c7d3026997c4ace5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cb6e7cff762c845ff9c43e0b072eba7c653b02703f8c1f3a564000822e5af44\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5e5afe4e2621897256e5a3c88f4d62db565155cf11abdf0c0de27b25ae80c03d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e578c3d1b3899861d7c1a717a90666abc21e6fc257e211f31c41cd2b966f239\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:04:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:04:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhp4b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ppmcr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.354337 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-4bcxm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc086f6b-af67-49e4-97c8-f8b70f19e49a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f10de997ad103241d497848f2236116ef34a903e35825e3d55f4a587c040a577\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gzmhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:57Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-4bcxm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.400294 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fcc1e5b-d0aa-4b28-ab66-3bbdc465b321\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3e954f4e5a3227b78bdbcf9adfd78ae7587f4edcd2d7eba76da5dcf3e8a0c5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://244c5e1c2187811c5b1b53000c2ac14fb97aaa7e1479925bebbe1a5aa154831f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://687f0acf340276893b88aed014fefe90fd67168d72bd0e19af64840356261e8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2de4a5c6ab353cee1b62449520d5300bc915036e53ab7a40be7c8f80e2264753\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4119ad5bdb9f5da977cc828824b870ee85b49100185d693c72c91d4b5f8d0b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89555cc4f831d0a25b05003527780e3a2d285fdda064190c2a50afcb1bccbdd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e89aa3db6cf48560fcfbbff3e2de953d0d1fe65f6ae638ecc10bc8251cb445fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cd1e669a65eace389a63d22dfc020461db7f58ad0e9f3e51f618d930d762ace9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.416767 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0a1030649d4c0733154c0864f97c0b26d129607d34d94a65ba69a9084f04d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.429849 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.429882 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.429894 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.429911 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.429923 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.438555 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87e68d183ad891d28a3bac2ae2e2b2f878b3c1a708d657cc7e3111626157ae90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83f28a03a7763ed7498fc7272f1ed4ab6be13b2ff941bc606b386a1cf7568d0a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.458943 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f036afbd-252a-4ed3-88e6-46256da87940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1880c78addfa5865cfdb73ac1d2965ff8142978ac0814615ea0d6ecb005f5847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4342a165126bd52a03ab2a8ac09666d08d16d3b8034de7b6be1ef02506798c94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef7ee0b4dfd54ec0a33df18eba05dbd234ef0ed39fe66b05ee5d8254614955fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e8c5cccdbbac0a338bd0c05acb0ffb20179ca7413df27cc499ab2fbfc9451d51\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.473247 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.490102 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.507955 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-x57xf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b88e9d2e-35da-45a8-ac7e-22afd660ff9f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-24T00:04:42Z\\\",\\\"message\\\":\\\"2026-01-24T00:03:57+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf\\\\n2026-01-24T00:03:57+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c1d5c5ee-6cad-41d4-9283-93b019ff77bf to /host/opt/cni/bin/\\\\n2026-01-24T00:03:57Z [verbose] multus-daemon started\\\\n2026-01-24T00:03:57Z [verbose] Readiness Indicator file check\\\\n2026-01-24T00:04:42Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:04:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-67bbh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-x57xf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.524607 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd647b0d-6d3d-432d-81ac-6484a2948211\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10947bd839858a55b7b098d2a83f3539d2000c9e32bef961d1e3b418516afbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w9vrg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7mzrz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.532660 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.532756 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.532805 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.532836 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.532854 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.539794 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fad2d146-bbd6-4720-8eec-7644ef2c0855\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://04c7475f668593c7097dbf2dd1453baa25cff2333367eadc62a1124a240dfe05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c8e40d7ba93858915cd80a18bfee202c1e6b2672cd41eff2441d6d5178d98e1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e40d7ba93858915cd80a18bfee202c1e6b2672cd41eff2441d6d5178d98e1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.560544 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"653e6c74-9f8e-4c5f-b101-5b8da2e962ca\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-24T00:03:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0124 00:03:48.618772 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0124 00:03:48.623114 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1018791735/tls.crt::/tmp/serving-cert-1018791735/tls.key\\\\\\\"\\\\nI0124 00:03:54.397485 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0124 00:03:54.405962 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0124 00:03:54.405983 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0124 00:03:54.406004 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0124 00:03:54.406008 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0124 00:03:54.413619 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0124 00:03:54.413647 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0124 00:03:54.413654 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0124 00:03:54.413652 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0124 00:03:54.413659 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0124 00:03:54.413676 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0124 00:03:54.413680 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0124 00:03:54.413684 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0124 00:03:54.415845 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-24T00:03:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.579852 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ef6c70c-58da-4218-be4c-8a1d15f72b06\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f74bb1b0407748b9f3b691a7fad9b13b58992e3688169fda4422379be523ab02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9b4a29a22859cdb13f508e7fbc10d00784a61df558cb6cb84079591e7184bf7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cca67c8957fa68d0167bb1892013f5a9447528a241a81c7b0626e256454edd75\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:36Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.595670 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-r4q22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18335446-e572-4741-ad9e-e7aadee7550b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:04:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tsw85\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:04:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-r4q22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.614450 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:57Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c741baa2e67719ad071fb354e213c74b40c67ff72c8c96ec612148344f07413b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.631193 4676 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5dg9q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"efe79b06-a59d-4d3c-9161-839d4e60fb52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-24T00:03:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53fef8a199be8ea38c412591af86a6bd9b703bce2a0662a8a61b10ffcb42b17b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-24T00:03:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cht5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-24T00:03:55Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5dg9q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:06Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.635203 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.635253 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.635270 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.635295 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.635313 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.738364 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.738451 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.738470 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.738495 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.738515 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.840856 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.840913 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.840929 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.840963 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.840980 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.943432 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.943498 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.943516 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.943540 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:06 crc kubenswrapper[4676]: I0124 00:05:06.943559 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:06Z","lastTransitionTime":"2026-01-24T00:05:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.046162 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.046231 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.046254 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.046281 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.046298 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.149075 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.149125 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.149144 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.149344 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.149446 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.252342 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.252437 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.252455 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.252479 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.252498 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.255089 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.255127 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:07 crc kubenswrapper[4676]: E0124 00:05:07.255238 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:07 crc kubenswrapper[4676]: E0124 00:05:07.255341 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.310253 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 05:41:19.374833952 +0000 UTC Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.355024 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.355083 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.355106 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.355133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.355154 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.458357 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.458600 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.458626 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.458651 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.458668 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.561775 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.561868 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.561893 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.561926 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.561950 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.665134 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.665215 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.665233 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.665265 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.665289 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.767801 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.767883 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.767907 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.767940 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.767963 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.870817 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.870875 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.870892 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.870915 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.870931 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.973901 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.973973 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.973997 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.974028 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:07 crc kubenswrapper[4676]: I0124 00:05:07.974050 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:07Z","lastTransitionTime":"2026-01-24T00:05:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.076947 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.076978 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.076988 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.077003 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.077014 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:08Z","lastTransitionTime":"2026-01-24T00:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.179339 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.179452 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.179524 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.179589 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.179599 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:08Z","lastTransitionTime":"2026-01-24T00:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.255539 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.255572 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:08 crc kubenswrapper[4676]: E0124 00:05:08.255706 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:08 crc kubenswrapper[4676]: E0124 00:05:08.255821 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.281671 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.281697 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.281705 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.281716 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.281725 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:08Z","lastTransitionTime":"2026-01-24T00:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.311416 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 12:31:18.427839008 +0000 UTC Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.384299 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.384356 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.384423 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.384455 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.384475 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:08Z","lastTransitionTime":"2026-01-24T00:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.487571 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.487618 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.487630 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.487648 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.487658 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:08Z","lastTransitionTime":"2026-01-24T00:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.590589 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.590643 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.590654 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.590675 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.590689 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:08Z","lastTransitionTime":"2026-01-24T00:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.693724 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.693782 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.693799 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.693825 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.693842 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:08Z","lastTransitionTime":"2026-01-24T00:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.796664 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.796745 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.796769 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.796794 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.796811 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:08Z","lastTransitionTime":"2026-01-24T00:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.899553 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.899616 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.899634 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.899657 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:08 crc kubenswrapper[4676]: I0124 00:05:08.899674 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:08Z","lastTransitionTime":"2026-01-24T00:05:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.002292 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.002475 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.002498 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.002525 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.002542 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.013631 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.013863 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.014114 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.014335 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.014582 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: E0124 00:05:09.045867 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.051446 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.051780 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.051980 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.052166 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.052326 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: E0124 00:05:09.073484 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.080546 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.080804 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.080913 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.081026 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.081128 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: E0124 00:05:09.105668 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.108964 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.109009 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.109026 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.109046 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.109063 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: E0124 00:05:09.123876 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.127122 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.127162 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.127177 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.127198 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.127216 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: E0124 00:05:09.140562 4676 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-24T00:05:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"55c3ff0e-ee2f-473a-9424-ac0aeb395b03\\\",\\\"systemUUID\\\":\\\"d7308ad2-105f-4282-b3b4-bf5b6bfb52ce\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-24T00:05:09Z is after 2025-08-24T17:21:41Z" Jan 24 00:05:09 crc kubenswrapper[4676]: E0124 00:05:09.140779 4676 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.142790 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.142876 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.142939 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.142999 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.143074 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.245885 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.246210 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.246459 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.246696 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.246859 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.255468 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:09 crc kubenswrapper[4676]: E0124 00:05:09.255606 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.255822 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:09 crc kubenswrapper[4676]: E0124 00:05:09.255995 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.312024 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 10:50:20.757710369 +0000 UTC Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.350034 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.350297 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.350423 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.350557 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.350738 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.453753 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.453816 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.453838 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.453867 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.453887 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.556545 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.556596 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.556613 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.556657 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.556673 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.659818 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.660138 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.660273 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.660452 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.660600 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.763878 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.763937 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.763955 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.763979 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.763997 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.866558 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.866587 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.866598 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.866614 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.866625 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.969657 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.969713 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.969734 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.969760 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:09 crc kubenswrapper[4676]: I0124 00:05:09.969779 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:09Z","lastTransitionTime":"2026-01-24T00:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.072324 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.072401 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.072419 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.072446 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.072463 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:10Z","lastTransitionTime":"2026-01-24T00:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.174558 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.174603 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.174619 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.174640 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.174657 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:10Z","lastTransitionTime":"2026-01-24T00:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.255355 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:10 crc kubenswrapper[4676]: E0124 00:05:10.255602 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.255904 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:10 crc kubenswrapper[4676]: E0124 00:05:10.256012 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.277526 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.277578 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.277589 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.277606 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.277620 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:10Z","lastTransitionTime":"2026-01-24T00:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.312895 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 00:05:29.951293816 +0000 UTC Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.381095 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.381164 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.381182 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.381207 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.381226 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:10Z","lastTransitionTime":"2026-01-24T00:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.483929 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.483993 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.484010 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.484034 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.484051 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:10Z","lastTransitionTime":"2026-01-24T00:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.587577 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.587643 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.587668 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.587717 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.587740 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:10Z","lastTransitionTime":"2026-01-24T00:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.690937 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.690999 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.691017 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.691044 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.691061 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:10Z","lastTransitionTime":"2026-01-24T00:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.793997 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.794088 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.794106 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.794166 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.794185 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:10Z","lastTransitionTime":"2026-01-24T00:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.897103 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.897161 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.897180 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.897204 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:10 crc kubenswrapper[4676]: I0124 00:05:10.897221 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:10Z","lastTransitionTime":"2026-01-24T00:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.001183 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.001253 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.001265 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.001282 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.001304 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.103997 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.104049 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.104065 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.104089 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.104106 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.207521 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.207573 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.207584 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.207601 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.207615 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.255098 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.255153 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:11 crc kubenswrapper[4676]: E0124 00:05:11.255287 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:11 crc kubenswrapper[4676]: E0124 00:05:11.255430 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.311593 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.311695 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.311751 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.311776 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.311797 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.313881 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 07:09:59.620620673 +0000 UTC Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.415808 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.415859 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.415875 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.415902 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.415920 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.518985 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.519037 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.519059 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.519087 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.519108 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.623411 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.623761 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.623780 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.623804 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.623821 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.726416 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.726483 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.726505 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.726534 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.726556 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.829249 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.829329 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.829341 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.829364 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.829410 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.933834 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.933889 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.933905 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.933932 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:11 crc kubenswrapper[4676]: I0124 00:05:11.933950 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:11Z","lastTransitionTime":"2026-01-24T00:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.037356 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.037439 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.037455 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.037478 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.037495 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.140193 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.140252 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.140268 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.140291 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.140307 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.242753 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.242841 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.242864 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.242895 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.242914 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.255130 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.255240 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:12 crc kubenswrapper[4676]: E0124 00:05:12.255344 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:12 crc kubenswrapper[4676]: E0124 00:05:12.255567 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.314023 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 06:30:14.928363648 +0000 UTC Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.345575 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.345642 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.345660 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.345685 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.345703 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.449425 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.449496 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.449518 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.449545 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.449566 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.552610 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.552668 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.552685 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.552708 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.552729 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.656275 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.656338 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.656356 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.656413 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.656434 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.759731 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.759810 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.759862 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.759893 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.759915 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.862485 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.862549 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.862568 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.862592 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.862609 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.975030 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.975098 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.975119 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.975148 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:12 crc kubenswrapper[4676]: I0124 00:05:12.975169 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:12Z","lastTransitionTime":"2026-01-24T00:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.078292 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.078415 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.078452 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.078483 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.078505 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:13Z","lastTransitionTime":"2026-01-24T00:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.180832 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.180880 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.180896 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.180921 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.180939 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:13Z","lastTransitionTime":"2026-01-24T00:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.255184 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.255204 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:13 crc kubenswrapper[4676]: E0124 00:05:13.255360 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:13 crc kubenswrapper[4676]: E0124 00:05:13.255537 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.256665 4676 scope.go:117] "RemoveContainer" containerID="dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9" Jan 24 00:05:13 crc kubenswrapper[4676]: E0124 00:05:13.256949 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-ld569_openshift-ovn-kubernetes(24f0dc26-0857-430f-aebd-073fcfcc1c0a)\"" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.283305 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.283354 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.283371 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.283416 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.283435 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:13Z","lastTransitionTime":"2026-01-24T00:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.314993 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 00:50:48.041864878 +0000 UTC Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.385941 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.385990 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.386000 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.386017 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.386028 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:13Z","lastTransitionTime":"2026-01-24T00:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.487958 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.488003 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.488014 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.488030 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.488040 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:13Z","lastTransitionTime":"2026-01-24T00:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.590587 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.590645 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.590661 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.590684 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.590703 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:13Z","lastTransitionTime":"2026-01-24T00:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.693930 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.694092 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.694118 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.694186 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.694203 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:13Z","lastTransitionTime":"2026-01-24T00:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.797340 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.797398 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.797409 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.797423 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.797433 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:13Z","lastTransitionTime":"2026-01-24T00:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.899945 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.900005 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.900027 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.900056 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:13 crc kubenswrapper[4676]: I0124 00:05:13.900077 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:13Z","lastTransitionTime":"2026-01-24T00:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.002798 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.002833 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.002842 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.002857 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.002868 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.105835 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.105875 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.105887 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.105902 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.105914 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.209805 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.209866 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.209884 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.209912 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.209930 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.255777 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.255884 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:14 crc kubenswrapper[4676]: E0124 00:05:14.256275 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:14 crc kubenswrapper[4676]: E0124 00:05:14.256546 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.313030 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.313103 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.313122 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.313147 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.313164 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.315119 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 19:05:29.76866111 +0000 UTC Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.415975 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.416039 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.416065 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.416094 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.416115 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.419729 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:14 crc kubenswrapper[4676]: E0124 00:05:14.419949 4676 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:05:14 crc kubenswrapper[4676]: E0124 00:05:14.420051 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs podName:18335446-e572-4741-ad9e-e7aadee7550b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:18.420024754 +0000 UTC m=+162.449995785 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs") pod "network-metrics-daemon-r4q22" (UID: "18335446-e572-4741-ad9e-e7aadee7550b") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.519369 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.519456 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.519473 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.519501 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.519518 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.623047 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.623092 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.623108 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.623131 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.623147 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.726042 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.726102 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.726119 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.726141 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.726158 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.828989 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.829053 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.829079 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.829104 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.829121 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.932756 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.932855 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.932881 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.932914 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:14 crc kubenswrapper[4676]: I0124 00:05:14.932933 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:14Z","lastTransitionTime":"2026-01-24T00:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.035788 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.035828 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.035837 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.035907 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.035919 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.139567 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.139645 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.139669 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.139699 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.139722 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.242674 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.242746 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.242764 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.242789 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.242810 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.255022 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.255074 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:15 crc kubenswrapper[4676]: E0124 00:05:15.255270 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:15 crc kubenswrapper[4676]: E0124 00:05:15.255471 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.315435 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 04:58:38.852511166 +0000 UTC Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.345431 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.345503 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.345524 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.345607 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.345626 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.448805 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.448857 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.448873 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.448897 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.448911 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.551770 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.551829 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.551846 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.551870 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.551886 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.654806 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.654877 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.654896 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.654922 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.654943 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.758040 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.758106 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.758126 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.758156 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.758172 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.861648 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.861684 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.861692 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.861705 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.861714 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.963891 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.963947 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.963964 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.963989 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:15 crc kubenswrapper[4676]: I0124 00:05:15.964007 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:15Z","lastTransitionTime":"2026-01-24T00:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.067725 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.067895 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.067967 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.068004 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.068074 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.171645 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.171722 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.171745 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.171777 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.171803 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.255287 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.255309 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:16 crc kubenswrapper[4676]: E0124 00:05:16.255492 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:16 crc kubenswrapper[4676]: E0124 00:05:16.255620 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.275107 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.275179 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.275196 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.275221 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.275238 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.315750 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 11:22:25.56358743 +0000 UTC Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.341734 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5dg9q" podStartSLOduration=81.341680189 podStartE2EDuration="1m21.341680189s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.298718726 +0000 UTC m=+100.328689787" watchObservedRunningTime="2026-01-24 00:05:16.341680189 +0000 UTC m=+100.371651230" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.360711 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7m8ts" podStartSLOduration=80.360687553 podStartE2EDuration="1m20.360687553s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.359340159 +0000 UTC m=+100.389311200" watchObservedRunningTime="2026-01-24 00:05:16.360687553 +0000 UTC m=+100.390658584" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.377552 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.377594 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.377611 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.377633 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.377649 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.451578 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-4bcxm" podStartSLOduration=81.45154997 podStartE2EDuration="1m21.45154997s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.45124015 +0000 UTC m=+100.481211181" watchObservedRunningTime="2026-01-24 00:05:16.45154997 +0000 UTC m=+100.481521011" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.452095 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ppmcr" podStartSLOduration=81.452085347 podStartE2EDuration="1m21.452085347s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.437223648 +0000 UTC m=+100.467194689" watchObservedRunningTime="2026-01-24 00:05:16.452085347 +0000 UTC m=+100.482056388" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.480278 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.480330 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.480343 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.480360 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.480373 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.494894 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=78.494875016 podStartE2EDuration="1m18.494875016s" podCreationTimestamp="2026-01-24 00:03:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.494799954 +0000 UTC m=+100.524770995" watchObservedRunningTime="2026-01-24 00:05:16.494875016 +0000 UTC m=+100.524846007" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.540904 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=80.540875358 podStartE2EDuration="1m20.540875358s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.53998465 +0000 UTC m=+100.569955661" watchObservedRunningTime="2026-01-24 00:05:16.540875358 +0000 UTC m=+100.570846369" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.557929 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=52.557899776 podStartE2EDuration="52.557899776s" podCreationTimestamp="2026-01-24 00:04:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.557076709 +0000 UTC m=+100.587047730" watchObservedRunningTime="2026-01-24 00:05:16.557899776 +0000 UTC m=+100.587870787" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.585807 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.585858 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.585871 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.585891 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.585905 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.617004 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-x57xf" podStartSLOduration=81.61698195 podStartE2EDuration="1m21.61698195s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.600676585 +0000 UTC m=+100.630647586" watchObservedRunningTime="2026-01-24 00:05:16.61698195 +0000 UTC m=+100.646952951" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.618190 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podStartSLOduration=81.618164138 podStartE2EDuration="1m21.618164138s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.616580797 +0000 UTC m=+100.646551798" watchObservedRunningTime="2026-01-24 00:05:16.618164138 +0000 UTC m=+100.648135139" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.631266 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=17.63124431 podStartE2EDuration="17.63124431s" podCreationTimestamp="2026-01-24 00:04:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.630737873 +0000 UTC m=+100.660708884" watchObservedRunningTime="2026-01-24 00:05:16.63124431 +0000 UTC m=+100.661215311" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.664648 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=81.664626205 podStartE2EDuration="1m21.664626205s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:16.649863409 +0000 UTC m=+100.679834430" watchObservedRunningTime="2026-01-24 00:05:16.664626205 +0000 UTC m=+100.694597216" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.687983 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.688046 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.688058 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.688075 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.688088 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.791137 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.791218 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.791237 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.791264 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.791282 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.893919 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.893973 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.893991 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.894055 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.894073 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.997051 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.997099 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.997115 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.997133 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:16 crc kubenswrapper[4676]: I0124 00:05:16.997145 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:16Z","lastTransitionTime":"2026-01-24T00:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.099993 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.100034 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.100046 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.100064 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.100075 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:17Z","lastTransitionTime":"2026-01-24T00:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.203514 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.203564 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.203581 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.203602 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.203629 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:17Z","lastTransitionTime":"2026-01-24T00:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.255295 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.255295 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:17 crc kubenswrapper[4676]: E0124 00:05:17.255732 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:17 crc kubenswrapper[4676]: E0124 00:05:17.255559 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.306930 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.306979 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.306996 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.307018 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.307035 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:17Z","lastTransitionTime":"2026-01-24T00:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.316827 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 03:34:45.480449053 +0000 UTC Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.410914 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.410996 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.411024 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.411525 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.411553 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:17Z","lastTransitionTime":"2026-01-24T00:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.514843 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.514914 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.514942 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.514970 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.514992 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:17Z","lastTransitionTime":"2026-01-24T00:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.617848 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.617893 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.617909 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.617932 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.617949 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:17Z","lastTransitionTime":"2026-01-24T00:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.721119 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.721178 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.721197 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.721219 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.721235 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:17Z","lastTransitionTime":"2026-01-24T00:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.823937 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.824299 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.824464 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.824621 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.824756 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:17Z","lastTransitionTime":"2026-01-24T00:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.927830 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.927895 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.927914 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.927939 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:17 crc kubenswrapper[4676]: I0124 00:05:17.927956 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:17Z","lastTransitionTime":"2026-01-24T00:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.031153 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.031213 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.031237 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.031268 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.031292 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.133725 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.133788 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.133810 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.133839 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.133865 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.236030 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.236098 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.236119 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.236144 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.236160 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.254685 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.254763 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:18 crc kubenswrapper[4676]: E0124 00:05:18.254850 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:18 crc kubenswrapper[4676]: E0124 00:05:18.254979 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.317530 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 23:45:07.92895782 +0000 UTC Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.338345 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.338406 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.338417 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.338435 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.338444 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.441654 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.441692 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.441704 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.441719 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.441730 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.544907 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.544973 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.544996 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.545026 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.545047 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.648428 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.648481 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.648498 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.648546 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.648565 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.751690 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.751751 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.751768 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.751796 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.751817 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.854878 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.854951 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.854970 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.854998 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.855018 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.957537 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.957591 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.957608 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.957632 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:18 crc kubenswrapper[4676]: I0124 00:05:18.957649 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:18Z","lastTransitionTime":"2026-01-24T00:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.060428 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.060556 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.060586 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.060644 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.060668 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:19Z","lastTransitionTime":"2026-01-24T00:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.163334 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.163429 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.163453 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.163482 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.163502 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:19Z","lastTransitionTime":"2026-01-24T00:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.254752 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.254797 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:19 crc kubenswrapper[4676]: E0124 00:05:19.254864 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:19 crc kubenswrapper[4676]: E0124 00:05:19.254996 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.266431 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.266494 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.266511 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.266535 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.266553 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:19Z","lastTransitionTime":"2026-01-24T00:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.317868 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 13:23:19.262423048 +0000 UTC Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.368136 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.368178 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.368188 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.368202 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.368214 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:19Z","lastTransitionTime":"2026-01-24T00:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.424370 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.424423 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.424431 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.424445 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.424457 4676 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-24T00:05:19Z","lastTransitionTime":"2026-01-24T00:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.488510 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t"] Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.489019 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.495047 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.495356 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.495607 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.495861 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.573630 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9c661b89-85ef-4810-9f35-7a3157d92f19-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.573729 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c661b89-85ef-4810-9f35-7a3157d92f19-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.573785 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c661b89-85ef-4810-9f35-7a3157d92f19-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.573981 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9c661b89-85ef-4810-9f35-7a3157d92f19-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.574042 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c661b89-85ef-4810-9f35-7a3157d92f19-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.676162 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9c661b89-85ef-4810-9f35-7a3157d92f19-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.676232 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c661b89-85ef-4810-9f35-7a3157d92f19-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.676286 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c661b89-85ef-4810-9f35-7a3157d92f19-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.676296 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9c661b89-85ef-4810-9f35-7a3157d92f19-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.676341 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c661b89-85ef-4810-9f35-7a3157d92f19-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.676432 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9c661b89-85ef-4810-9f35-7a3157d92f19-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.676620 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9c661b89-85ef-4810-9f35-7a3157d92f19-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.678618 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9c661b89-85ef-4810-9f35-7a3157d92f19-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.684801 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9c661b89-85ef-4810-9f35-7a3157d92f19-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.711295 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c661b89-85ef-4810-9f35-7a3157d92f19-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mfp4t\" (UID: \"9c661b89-85ef-4810-9f35-7a3157d92f19\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.853902 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" Jan 24 00:05:19 crc kubenswrapper[4676]: I0124 00:05:19.960980 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" event={"ID":"9c661b89-85ef-4810-9f35-7a3157d92f19","Type":"ContainerStarted","Data":"6b81bde11a23224074d6a8a11874f6f0e34b2aac36339b4d8136f6f43fd02a68"} Jan 24 00:05:20 crc kubenswrapper[4676]: I0124 00:05:20.255802 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:20 crc kubenswrapper[4676]: E0124 00:05:20.256247 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:20 crc kubenswrapper[4676]: I0124 00:05:20.256627 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:20 crc kubenswrapper[4676]: E0124 00:05:20.256827 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:20 crc kubenswrapper[4676]: I0124 00:05:20.318754 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 21:28:22.979160938 +0000 UTC Jan 24 00:05:20 crc kubenswrapper[4676]: I0124 00:05:20.318887 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 24 00:05:20 crc kubenswrapper[4676]: I0124 00:05:20.330698 4676 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 24 00:05:20 crc kubenswrapper[4676]: I0124 00:05:20.966582 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" event={"ID":"9c661b89-85ef-4810-9f35-7a3157d92f19","Type":"ContainerStarted","Data":"d2bc90ef0b210390f6be3cbc85c806f3507e98e1817f4c761f79eb661e0e26e8"} Jan 24 00:05:21 crc kubenswrapper[4676]: I0124 00:05:21.255575 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:21 crc kubenswrapper[4676]: E0124 00:05:21.255733 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:21 crc kubenswrapper[4676]: I0124 00:05:21.255577 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:21 crc kubenswrapper[4676]: E0124 00:05:21.256027 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:22 crc kubenswrapper[4676]: I0124 00:05:22.254909 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:22 crc kubenswrapper[4676]: I0124 00:05:22.254940 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:22 crc kubenswrapper[4676]: E0124 00:05:22.256010 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:22 crc kubenswrapper[4676]: E0124 00:05:22.256569 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:23 crc kubenswrapper[4676]: I0124 00:05:23.255510 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:23 crc kubenswrapper[4676]: I0124 00:05:23.255568 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:23 crc kubenswrapper[4676]: E0124 00:05:23.255682 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:23 crc kubenswrapper[4676]: E0124 00:05:23.255790 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:24 crc kubenswrapper[4676]: I0124 00:05:24.255051 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:24 crc kubenswrapper[4676]: I0124 00:05:24.255083 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:24 crc kubenswrapper[4676]: E0124 00:05:24.255229 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:24 crc kubenswrapper[4676]: E0124 00:05:24.255363 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:25 crc kubenswrapper[4676]: I0124 00:05:25.255341 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:25 crc kubenswrapper[4676]: I0124 00:05:25.255430 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:25 crc kubenswrapper[4676]: E0124 00:05:25.255560 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:25 crc kubenswrapper[4676]: E0124 00:05:25.255693 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:26 crc kubenswrapper[4676]: I0124 00:05:26.255681 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:26 crc kubenswrapper[4676]: I0124 00:05:26.255681 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:26 crc kubenswrapper[4676]: E0124 00:05:26.257997 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:26 crc kubenswrapper[4676]: E0124 00:05:26.257722 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:27 crc kubenswrapper[4676]: I0124 00:05:27.255573 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:27 crc kubenswrapper[4676]: I0124 00:05:27.255827 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:27 crc kubenswrapper[4676]: E0124 00:05:27.255946 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:27 crc kubenswrapper[4676]: E0124 00:05:27.256110 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:28 crc kubenswrapper[4676]: I0124 00:05:28.255605 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:28 crc kubenswrapper[4676]: E0124 00:05:28.256090 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:28 crc kubenswrapper[4676]: I0124 00:05:28.256325 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:28 crc kubenswrapper[4676]: I0124 00:05:28.256522 4676 scope.go:117] "RemoveContainer" containerID="dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9" Jan 24 00:05:28 crc kubenswrapper[4676]: E0124 00:05:28.256651 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:28 crc kubenswrapper[4676]: I0124 00:05:28.992849 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/1.log" Jan 24 00:05:28 crc kubenswrapper[4676]: I0124 00:05:28.993234 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/0.log" Jan 24 00:05:28 crc kubenswrapper[4676]: I0124 00:05:28.993262 4676 generic.go:334] "Generic (PLEG): container finished" podID="b88e9d2e-35da-45a8-ac7e-22afd660ff9f" containerID="cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b" exitCode=1 Jan 24 00:05:28 crc kubenswrapper[4676]: I0124 00:05:28.993306 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x57xf" event={"ID":"b88e9d2e-35da-45a8-ac7e-22afd660ff9f","Type":"ContainerDied","Data":"cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b"} Jan 24 00:05:28 crc kubenswrapper[4676]: I0124 00:05:28.993340 4676 scope.go:117] "RemoveContainer" containerID="db43410c7c6a0f160ce59403dc22a9b216d73ef62bebd77daf8f6e6818ed733c" Jan 24 00:05:28 crc kubenswrapper[4676]: I0124 00:05:28.993691 4676 scope.go:117] "RemoveContainer" containerID="cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b" Jan 24 00:05:28 crc kubenswrapper[4676]: E0124 00:05:28.993824 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-x57xf_openshift-multus(b88e9d2e-35da-45a8-ac7e-22afd660ff9f)\"" pod="openshift-multus/multus-x57xf" podUID="b88e9d2e-35da-45a8-ac7e-22afd660ff9f" Jan 24 00:05:29 crc kubenswrapper[4676]: I0124 00:05:29.003105 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/3.log" Jan 24 00:05:29 crc kubenswrapper[4676]: I0124 00:05:29.007538 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerStarted","Data":"b097984a7c57abb09d2fc1362781982b54bf19474644ee1742c87013905c7faf"} Jan 24 00:05:29 crc kubenswrapper[4676]: I0124 00:05:29.008354 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:05:29 crc kubenswrapper[4676]: I0124 00:05:29.021801 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mfp4t" podStartSLOduration=94.021786014 podStartE2EDuration="1m34.021786014s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:20.988732808 +0000 UTC m=+105.018703849" watchObservedRunningTime="2026-01-24 00:05:29.021786014 +0000 UTC m=+113.051757025" Jan 24 00:05:29 crc kubenswrapper[4676]: I0124 00:05:29.060609 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podStartSLOduration=94.060590084 podStartE2EDuration="1m34.060590084s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:29.059270512 +0000 UTC m=+113.089241523" watchObservedRunningTime="2026-01-24 00:05:29.060590084 +0000 UTC m=+113.090561085" Jan 24 00:05:29 crc kubenswrapper[4676]: I0124 00:05:29.250020 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-r4q22"] Jan 24 00:05:29 crc kubenswrapper[4676]: I0124 00:05:29.250182 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:29 crc kubenswrapper[4676]: E0124 00:05:29.250320 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:29 crc kubenswrapper[4676]: I0124 00:05:29.254959 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:29 crc kubenswrapper[4676]: E0124 00:05:29.255125 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:29 crc kubenswrapper[4676]: I0124 00:05:29.255494 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:29 crc kubenswrapper[4676]: E0124 00:05:29.255666 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:30 crc kubenswrapper[4676]: I0124 00:05:30.013612 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/1.log" Jan 24 00:05:30 crc kubenswrapper[4676]: I0124 00:05:30.254982 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:30 crc kubenswrapper[4676]: E0124 00:05:30.255164 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:31 crc kubenswrapper[4676]: I0124 00:05:31.254818 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:31 crc kubenswrapper[4676]: I0124 00:05:31.254918 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:31 crc kubenswrapper[4676]: I0124 00:05:31.254919 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:31 crc kubenswrapper[4676]: E0124 00:05:31.255017 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:31 crc kubenswrapper[4676]: E0124 00:05:31.255160 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:31 crc kubenswrapper[4676]: E0124 00:05:31.255267 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:32 crc kubenswrapper[4676]: I0124 00:05:32.254864 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:32 crc kubenswrapper[4676]: E0124 00:05:32.255044 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:33 crc kubenswrapper[4676]: I0124 00:05:33.254898 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:33 crc kubenswrapper[4676]: I0124 00:05:33.254907 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:33 crc kubenswrapper[4676]: I0124 00:05:33.254902 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:33 crc kubenswrapper[4676]: E0124 00:05:33.255269 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:33 crc kubenswrapper[4676]: E0124 00:05:33.255350 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:33 crc kubenswrapper[4676]: E0124 00:05:33.255045 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:34 crc kubenswrapper[4676]: I0124 00:05:34.255713 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:34 crc kubenswrapper[4676]: E0124 00:05:34.255996 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:35 crc kubenswrapper[4676]: I0124 00:05:35.255489 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:35 crc kubenswrapper[4676]: I0124 00:05:35.255584 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:35 crc kubenswrapper[4676]: I0124 00:05:35.255519 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:35 crc kubenswrapper[4676]: E0124 00:05:35.255679 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:35 crc kubenswrapper[4676]: E0124 00:05:35.255808 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:35 crc kubenswrapper[4676]: E0124 00:05:35.256002 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:36 crc kubenswrapper[4676]: E0124 00:05:36.199413 4676 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 24 00:05:36 crc kubenswrapper[4676]: I0124 00:05:36.254901 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:36 crc kubenswrapper[4676]: E0124 00:05:36.256081 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:36 crc kubenswrapper[4676]: E0124 00:05:36.364373 4676 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 24 00:05:37 crc kubenswrapper[4676]: I0124 00:05:37.255009 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:37 crc kubenswrapper[4676]: E0124 00:05:37.255139 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:37 crc kubenswrapper[4676]: I0124 00:05:37.255017 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:37 crc kubenswrapper[4676]: I0124 00:05:37.255038 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:37 crc kubenswrapper[4676]: E0124 00:05:37.255275 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:37 crc kubenswrapper[4676]: E0124 00:05:37.255323 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:38 crc kubenswrapper[4676]: I0124 00:05:38.254818 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:38 crc kubenswrapper[4676]: E0124 00:05:38.254998 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:39 crc kubenswrapper[4676]: I0124 00:05:39.255759 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:39 crc kubenswrapper[4676]: I0124 00:05:39.255776 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:39 crc kubenswrapper[4676]: E0124 00:05:39.255977 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:39 crc kubenswrapper[4676]: I0124 00:05:39.255793 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:39 crc kubenswrapper[4676]: E0124 00:05:39.256133 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:39 crc kubenswrapper[4676]: E0124 00:05:39.256163 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:40 crc kubenswrapper[4676]: I0124 00:05:40.255273 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:40 crc kubenswrapper[4676]: E0124 00:05:40.255725 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:40 crc kubenswrapper[4676]: I0124 00:05:40.256036 4676 scope.go:117] "RemoveContainer" containerID="cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b" Jan 24 00:05:41 crc kubenswrapper[4676]: I0124 00:05:41.066685 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/1.log" Jan 24 00:05:41 crc kubenswrapper[4676]: I0124 00:05:41.067019 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x57xf" event={"ID":"b88e9d2e-35da-45a8-ac7e-22afd660ff9f","Type":"ContainerStarted","Data":"503448e193566525ada0f32c12c8a2978a0f18fbc763208a99e7e6534727cec5"} Jan 24 00:05:41 crc kubenswrapper[4676]: I0124 00:05:41.255663 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:41 crc kubenswrapper[4676]: I0124 00:05:41.255766 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:41 crc kubenswrapper[4676]: E0124 00:05:41.255822 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:41 crc kubenswrapper[4676]: I0124 00:05:41.255923 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:41 crc kubenswrapper[4676]: E0124 00:05:41.255959 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:41 crc kubenswrapper[4676]: E0124 00:05:41.256124 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:41 crc kubenswrapper[4676]: E0124 00:05:41.366168 4676 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 24 00:05:42 crc kubenswrapper[4676]: I0124 00:05:42.254865 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:42 crc kubenswrapper[4676]: E0124 00:05:42.255508 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:43 crc kubenswrapper[4676]: I0124 00:05:43.255451 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:43 crc kubenswrapper[4676]: I0124 00:05:43.255576 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:43 crc kubenswrapper[4676]: I0124 00:05:43.255474 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:43 crc kubenswrapper[4676]: E0124 00:05:43.255635 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:43 crc kubenswrapper[4676]: E0124 00:05:43.255777 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:43 crc kubenswrapper[4676]: E0124 00:05:43.255911 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:44 crc kubenswrapper[4676]: I0124 00:05:44.254894 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:44 crc kubenswrapper[4676]: E0124 00:05:44.255049 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:45 crc kubenswrapper[4676]: I0124 00:05:45.255485 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:45 crc kubenswrapper[4676]: I0124 00:05:45.255540 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:45 crc kubenswrapper[4676]: E0124 00:05:45.255605 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 24 00:05:45 crc kubenswrapper[4676]: I0124 00:05:45.255506 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:45 crc kubenswrapper[4676]: E0124 00:05:45.255694 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-r4q22" podUID="18335446-e572-4741-ad9e-e7aadee7550b" Jan 24 00:05:45 crc kubenswrapper[4676]: E0124 00:05:45.255833 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 24 00:05:46 crc kubenswrapper[4676]: I0124 00:05:46.255404 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:46 crc kubenswrapper[4676]: E0124 00:05:46.257086 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 24 00:05:47 crc kubenswrapper[4676]: I0124 00:05:47.255149 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:05:47 crc kubenswrapper[4676]: I0124 00:05:47.255191 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:05:47 crc kubenswrapper[4676]: I0124 00:05:47.255285 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:05:47 crc kubenswrapper[4676]: I0124 00:05:47.257966 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 24 00:05:47 crc kubenswrapper[4676]: I0124 00:05:47.258671 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 24 00:05:47 crc kubenswrapper[4676]: I0124 00:05:47.258929 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 24 00:05:47 crc kubenswrapper[4676]: I0124 00:05:47.259295 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 24 00:05:48 crc kubenswrapper[4676]: I0124 00:05:48.255146 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:05:48 crc kubenswrapper[4676]: I0124 00:05:48.257793 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 24 00:05:48 crc kubenswrapper[4676]: I0124 00:05:48.260457 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.202198 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.906569 4676 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.964151 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5"] Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.964624 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.965408 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qzmmz"] Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.965722 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.966243 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jmv6d"] Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.966642 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.967467 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2scbc"] Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.968089 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.968348 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.968804 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.969745 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.969948 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.970470 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.970519 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.970564 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.972156 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.974563 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.974632 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.974893 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.975166 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.979282 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.979719 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.980172 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.980337 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.980526 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.980784 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.980967 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.981647 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.982114 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.983081 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.983197 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.983458 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.980788 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.984811 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.991258 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.992893 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.993364 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 24 00:05:49 crc kubenswrapper[4676]: I0124 00:05:49.993584 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.003216 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7zllh"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.003884 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.004923 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-x649d"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.005410 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.006547 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.007748 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.008223 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kvcv8"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.011180 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.011327 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.011496 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.014444 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.016968 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.020015 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.020431 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.020830 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29486880-rhzd2"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.021200 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.021752 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.022304 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-g2smk"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.022902 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.027012 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.027173 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.027313 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.027552 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.032957 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.033249 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.034159 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.036853 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.037905 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.044432 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.048291 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.050919 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.061966 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.062392 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.064098 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.064226 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.064309 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.064396 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.064473 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ft4kq"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.064545 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.064615 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.065294 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.065349 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.065498 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.065643 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.065759 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.065900 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.066036 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.066232 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.066405 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.066455 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.066625 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.066408 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.066784 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.066859 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.067011 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.067187 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.068331 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.068817 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.068876 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.071426 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.073367 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.073510 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.074751 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.074881 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.075007 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.076506 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.076626 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.076727 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.076839 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.076909 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.077020 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.077116 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.084557 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6m9lm"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.084955 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.087946 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.088257 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.088473 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.088974 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.089123 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.089636 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090221 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b189330a-ee63-45f1-8104-4ef173f8ee22-audit-dir\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090268 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-oauth-serving-cert\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090301 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-etcd-client\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090324 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-trusted-ca-bundle\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090349 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-audit-dir\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090370 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35efd97c-0521-428a-896d-b67490207db5-serving-cert\") pod \"openshift-config-operator-7777fb866f-7zllh\" (UID: \"35efd97c-0521-428a-896d-b67490207db5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090479 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-client-ca\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090508 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-etcd-serving-ca\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090531 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-serving-cert\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090562 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p6c5\" (UniqueName: \"kubernetes.io/projected/35efd97c-0521-428a-896d-b67490207db5-kube-api-access-6p6c5\") pod \"openshift-config-operator-7777fb866f-7zllh\" (UID: \"35efd97c-0521-428a-896d-b67490207db5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090592 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-oauth-config\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090615 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6351c23-e315-4c92-a467-380da403d3c4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090642 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090666 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c51c973-c370-41e8-b167-25d3b11418bf-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090697 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qp87\" (UniqueName: \"kubernetes.io/projected/00beace7-1e83-40ed-8d92-6da0cae7817f-kube-api-access-4qp87\") pod \"image-pruner-29486880-rhzd2\" (UID: \"00beace7-1e83-40ed-8d92-6da0cae7817f\") " pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090723 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6360940-ea9d-456d-b546-5a20af404ee5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zc8ts\" (UID: \"a6360940-ea9d-456d-b546-5a20af404ee5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090748 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ksdz\" (UniqueName: \"kubernetes.io/projected/32de8698-4bd5-4154-92a3-76930504a72d-kube-api-access-4ksdz\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090777 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32de8698-4bd5-4154-92a3-76930504a72d-trusted-ca\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090804 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-service-ca\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090843 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/00beace7-1e83-40ed-8d92-6da0cae7817f-serviceca\") pod \"image-pruner-29486880-rhzd2\" (UID: \"00beace7-1e83-40ed-8d92-6da0cae7817f\") " pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090872 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-encryption-config\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090899 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b189330a-ee63-45f1-8104-4ef173f8ee22-serving-cert\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090927 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6351c23-e315-4c92-a467-380da403d3c4-serving-cert\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090954 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv9mz\" (UniqueName: \"kubernetes.io/projected/b37a2847-a94a-4a0c-b092-1ed7155a2d35-kube-api-access-nv9mz\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.090986 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hzh\" (UniqueName: \"kubernetes.io/projected/9c51c973-c370-41e8-b167-25d3b11418bf-kube-api-access-87hzh\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091011 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-audit\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091028 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6360940-ea9d-456d-b546-5a20af404ee5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zc8ts\" (UID: \"a6360940-ea9d-456d-b546-5a20af404ee5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091052 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/35efd97c-0521-428a-896d-b67490207db5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7zllh\" (UID: \"35efd97c-0521-428a-896d-b67490207db5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091092 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6dpw\" (UniqueName: \"kubernetes.io/projected/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-kube-api-access-f6dpw\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091116 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091144 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6351c23-e315-4c92-a467-380da403d3c4-service-ca-bundle\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091162 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091180 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-client-ca\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091198 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37a2847-a94a-4a0c-b092-1ed7155a2d35-serving-cert\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091214 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-config\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091233 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4cw9\" (UniqueName: \"kubernetes.io/projected/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-kube-api-access-h4cw9\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091251 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-serving-cert\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091296 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-policies\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091327 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091353 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091393 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-node-pullsecrets\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091413 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6351c23-e315-4c92-a467-380da403d3c4-config\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091430 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d8d73604-754b-4dea-9be4-3451964e5589-machine-approver-tls\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091455 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8911791-9db1-4463-997e-1ed50da17324-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091479 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-config\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091500 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqnsp\" (UniqueName: \"kubernetes.io/projected/a6360940-ea9d-456d-b546-5a20af404ee5-kube-api-access-jqnsp\") pod \"openshift-apiserver-operator-796bbdcf4f-zc8ts\" (UID: \"a6360940-ea9d-456d-b546-5a20af404ee5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091520 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8911791-9db1-4463-997e-1ed50da17324-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091543 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9c51c973-c370-41e8-b167-25d3b11418bf-images\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091566 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d8d73604-754b-4dea-9be4-3451964e5589-auth-proxy-config\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091591 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-image-import-ca\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091636 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b189330a-ee63-45f1-8104-4ef173f8ee22-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091680 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091709 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b189330a-ee63-45f1-8104-4ef173f8ee22-etcd-client\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091732 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c51c973-c370-41e8-b167-25d3b11418bf-config\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091754 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f9mm\" (UniqueName: \"kubernetes.io/projected/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-kube-api-access-5f9mm\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091778 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-config\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091800 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-serving-cert\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091827 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32de8698-4bd5-4154-92a3-76930504a72d-serving-cert\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091879 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091918 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmt92\" (UniqueName: \"kubernetes.io/projected/d8d73604-754b-4dea-9be4-3451964e5589-kube-api-access-cmt92\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091945 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b189330a-ee63-45f1-8104-4ef173f8ee22-audit-policies\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091969 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8911791-9db1-4463-997e-1ed50da17324-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.091997 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092008 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092022 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-config\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092048 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b189330a-ee63-45f1-8104-4ef173f8ee22-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092073 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8d73604-754b-4dea-9be4-3451964e5589-config\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092102 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fftp\" (UniqueName: \"kubernetes.io/projected/f8911791-9db1-4463-997e-1ed50da17324-kube-api-access-6fftp\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092129 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th8cd\" (UniqueName: \"kubernetes.io/projected/b189330a-ee63-45f1-8104-4ef173f8ee22-kube-api-access-th8cd\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092171 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092201 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32de8698-4bd5-4154-92a3-76930504a72d-config\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092252 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6vn7\" (UniqueName: \"kubernetes.io/projected/e6351c23-e315-4c92-a467-380da403d3c4-kube-api-access-q6vn7\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092277 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-dir\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.092302 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b189330a-ee63-45f1-8104-4ef173f8ee22-encryption-config\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.093966 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.094543 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wvvpg"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.094910 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6m9lm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.094923 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cdrbc"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.095026 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.095087 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.095656 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.095980 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.096320 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.097191 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.097492 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.097707 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.097926 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.111476 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.111725 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.111853 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.116787 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.117087 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.120519 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.121820 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.128481 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.129031 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ck624"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.129127 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.129307 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.129630 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.129911 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.130094 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.130452 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.131190 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.131454 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-zq745"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.132987 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.133334 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.133869 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.134367 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.135559 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.136572 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.136953 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6kx9j"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.138538 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.139027 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.141629 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.141961 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.142867 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.147813 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.148568 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.149086 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.149282 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.155460 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.156166 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.156799 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.156864 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.157153 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.160119 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rsw66"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.161038 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.168559 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.171083 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.172094 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.173735 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shf2w"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.174341 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.186093 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.186844 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.187443 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.190580 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.191217 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194626 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/00beace7-1e83-40ed-8d92-6da0cae7817f-serviceca\") pod \"image-pruner-29486880-rhzd2\" (UID: \"00beace7-1e83-40ed-8d92-6da0cae7817f\") " pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194654 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25db14a1-725f-42ae-a6e9-646546b584c7-metrics-tls\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194675 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlm94\" (UniqueName: \"kubernetes.io/projected/89435826-645d-48a2-aa3b-f5c42003dcbe-kube-api-access-tlm94\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194697 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk6nf\" (UniqueName: \"kubernetes.io/projected/d8eed212-9137-45e5-8347-1f921fbedb19-kube-api-access-nk6nf\") pod \"multus-admission-controller-857f4d67dd-ck624\" (UID: \"d8eed212-9137-45e5-8347-1f921fbedb19\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194717 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hzh\" (UniqueName: \"kubernetes.io/projected/9c51c973-c370-41e8-b167-25d3b11418bf-kube-api-access-87hzh\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194739 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/aca1e4a5-f702-4803-8f47-7fcb8c7326b6-srv-cert\") pod \"catalog-operator-68c6474976-2xb97\" (UID: \"aca1e4a5-f702-4803-8f47-7fcb8c7326b6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194761 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977bd2bf-e652-4b16-b8fc-902d4a1d7860-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nnjp\" (UID: \"977bd2bf-e652-4b16-b8fc-902d4a1d7860\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194801 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-audit\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194821 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194839 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194855 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7srgg\" (UniqueName: \"kubernetes.io/projected/977bd2bf-e652-4b16-b8fc-902d4a1d7860-kube-api-access-7srgg\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nnjp\" (UID: \"977bd2bf-e652-4b16-b8fc-902d4a1d7860\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194871 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-client-ca\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194886 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-apiservice-cert\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194905 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194921 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a35921c-db91-45c2-a92c-a42f5b2cab84-metrics-tls\") pod \"dns-operator-744455d44c-cdrbc\" (UID: \"8a35921c-db91-45c2-a92c-a42f5b2cab84\") " pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194939 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4bc9\" (UniqueName: \"kubernetes.io/projected/7b8f105b-569d-47f2-b564-a0830b010e31-kube-api-access-j4bc9\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194956 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-serving-cert\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194972 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-policies\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.194990 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c50241f8-d135-4df8-b047-e76fb28b8a3d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2cgk2\" (UID: \"c50241f8-d135-4df8-b047-e76fb28b8a3d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195010 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195026 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6351c23-e315-4c92-a467-380da403d3c4-config\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195042 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d8d73604-754b-4dea-9be4-3451964e5589-machine-approver-tls\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195060 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195079 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm4tl\" (UniqueName: \"kubernetes.io/projected/7af03157-ee92-4e72-a775-acaeabb73e65-kube-api-access-jm4tl\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195097 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6kdj\" (UniqueName: \"kubernetes.io/projected/9d605c02-a15d-46a8-942c-cd85e6ce5452-kube-api-access-h6kdj\") pod \"machine-config-controller-84d6567774-nhmlj\" (UID: \"9d605c02-a15d-46a8-942c-cd85e6ce5452\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195115 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4d53a745-f985-4659-b62c-ce297ce8ce85-signing-key\") pod \"service-ca-9c57cc56f-6kx9j\" (UID: \"4d53a745-f985-4659-b62c-ce297ce8ce85\") " pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195133 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqnsp\" (UniqueName: \"kubernetes.io/projected/a6360940-ea9d-456d-b546-5a20af404ee5-kube-api-access-jqnsp\") pod \"openshift-apiserver-operator-796bbdcf4f-zc8ts\" (UID: \"a6360940-ea9d-456d-b546-5a20af404ee5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195150 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9c51c973-c370-41e8-b167-25d3b11418bf-images\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195169 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d8d73604-754b-4dea-9be4-3451964e5589-auth-proxy-config\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195186 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2d89ff2-f919-459e-8089-5097aab0f4e2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sxjwn\" (UID: \"c2d89ff2-f919-459e-8089-5097aab0f4e2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195203 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-image-import-ca\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195222 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b189330a-ee63-45f1-8104-4ef173f8ee22-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195240 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195256 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9d605c02-a15d-46a8-942c-cd85e6ce5452-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nhmlj\" (UID: \"9d605c02-a15d-46a8-942c-cd85e6ce5452\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195272 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9350aae-3053-431e-a2d6-2137f990ca08-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9h9cg\" (UID: \"c9350aae-3053-431e-a2d6-2137f990ca08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195289 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c51c973-c370-41e8-b167-25d3b11418bf-config\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195305 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f9mm\" (UniqueName: \"kubernetes.io/projected/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-kube-api-access-5f9mm\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195322 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195337 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9350aae-3053-431e-a2d6-2137f990ca08-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9h9cg\" (UID: \"c9350aae-3053-431e-a2d6-2137f990ca08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195352 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d8eed212-9137-45e5-8347-1f921fbedb19-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ck624\" (UID: \"d8eed212-9137-45e5-8347-1f921fbedb19\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195368 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmt92\" (UniqueName: \"kubernetes.io/projected/d8d73604-754b-4dea-9be4-3451964e5589-kube-api-access-cmt92\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195403 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm9zj\" (UniqueName: \"kubernetes.io/projected/aca1e4a5-f702-4803-8f47-7fcb8c7326b6-kube-api-access-mm9zj\") pod \"catalog-operator-68c6474976-2xb97\" (UID: \"aca1e4a5-f702-4803-8f47-7fcb8c7326b6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195419 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195437 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-config\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195452 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3651bf5e-f692-42b1-8d5e-512daae90cc8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195467 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d89ff2-f919-459e-8089-5097aab0f4e2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sxjwn\" (UID: \"c2d89ff2-f919-459e-8089-5097aab0f4e2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195484 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8d73604-754b-4dea-9be4-3451964e5589-config\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195502 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96br9\" (UniqueName: \"kubernetes.io/projected/25db14a1-725f-42ae-a6e9-646546b584c7-kube-api-access-96br9\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195520 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th8cd\" (UniqueName: \"kubernetes.io/projected/b189330a-ee63-45f1-8104-4ef173f8ee22-kube-api-access-th8cd\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195546 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-dir\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195563 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-tmpfs\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195583 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-oauth-serving-cert\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195617 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7b8f105b-569d-47f2-b564-a0830b010e31-default-certificate\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195635 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/aca1e4a5-f702-4803-8f47-7fcb8c7326b6-profile-collector-cert\") pod \"catalog-operator-68c6474976-2xb97\" (UID: \"aca1e4a5-f702-4803-8f47-7fcb8c7326b6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195651 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gszgz\" (UniqueName: \"kubernetes.io/projected/c50241f8-d135-4df8-b047-e76fb28b8a3d-kube-api-access-gszgz\") pod \"olm-operator-6b444d44fb-2cgk2\" (UID: \"c50241f8-d135-4df8-b047-e76fb28b8a3d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195669 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p6c5\" (UniqueName: \"kubernetes.io/projected/35efd97c-0521-428a-896d-b67490207db5-kube-api-access-6p6c5\") pod \"openshift-config-operator-7777fb866f-7zllh\" (UID: \"35efd97c-0521-428a-896d-b67490207db5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195695 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-client-ca\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195727 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-etcd-serving-ca\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195751 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195768 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195783 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eabf562-d289-4685-8ee5-ed1525930d19-secret-volume\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195797 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-webhook-cert\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195813 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qp87\" (UniqueName: \"kubernetes.io/projected/00beace7-1e83-40ed-8d92-6da0cae7817f-kube-api-access-4qp87\") pod \"image-pruner-29486880-rhzd2\" (UID: \"00beace7-1e83-40ed-8d92-6da0cae7817f\") " pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195832 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6360940-ea9d-456d-b546-5a20af404ee5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zc8ts\" (UID: \"a6360940-ea9d-456d-b546-5a20af404ee5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195848 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ksdz\" (UniqueName: \"kubernetes.io/projected/32de8698-4bd5-4154-92a3-76930504a72d-kube-api-access-4ksdz\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195863 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-service-ca\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195879 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7b8f105b-569d-47f2-b564-a0830b010e31-stats-auth\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195897 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32de8698-4bd5-4154-92a3-76930504a72d-trusted-ca\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195912 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcwrj\" (UniqueName: \"kubernetes.io/projected/3651bf5e-f692-42b1-8d5e-512daae90cc8-kube-api-access-dcwrj\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195928 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b8f105b-569d-47f2-b564-a0830b010e31-metrics-certs\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195945 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv9mz\" (UniqueName: \"kubernetes.io/projected/b37a2847-a94a-4a0c-b092-1ed7155a2d35-kube-api-access-nv9mz\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195960 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-encryption-config\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195978 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b189330a-ee63-45f1-8104-4ef173f8ee22-serving-cert\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.195999 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6351c23-e315-4c92-a467-380da403d3c4-serving-cert\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196018 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-config\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196036 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196052 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6360940-ea9d-456d-b546-5a20af404ee5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zc8ts\" (UID: \"a6360940-ea9d-456d-b546-5a20af404ee5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196068 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/35efd97c-0521-428a-896d-b67490207db5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7zllh\" (UID: \"35efd97c-0521-428a-896d-b67490207db5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196085 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6dpw\" (UniqueName: \"kubernetes.io/projected/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-kube-api-access-f6dpw\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196102 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6351c23-e315-4c92-a467-380da403d3c4-service-ca-bundle\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196121 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196141 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/596a60f1-3eb7-4a57-af4a-1cb1f37f2824-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ljbnb\" (UID: \"596a60f1-3eb7-4a57-af4a-1cb1f37f2824\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196159 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxkrp\" (UniqueName: \"kubernetes.io/projected/4b605a66-7904-4596-a67f-ea21ef41a24b-kube-api-access-dxkrp\") pod \"downloads-7954f5f757-6m9lm\" (UID: \"4b605a66-7904-4596-a67f-ea21ef41a24b\") " pod="openshift-console/downloads-7954f5f757-6m9lm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196184 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4cw9\" (UniqueName: \"kubernetes.io/projected/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-kube-api-access-h4cw9\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196201 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37a2847-a94a-4a0c-b092-1ed7155a2d35-serving-cert\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196216 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-config\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196232 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-node-pullsecrets\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196251 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196267 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d89ff2-f919-459e-8089-5097aab0f4e2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sxjwn\" (UID: \"c2d89ff2-f919-459e-8089-5097aab0f4e2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196291 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8911791-9db1-4463-997e-1ed50da17324-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196306 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3651bf5e-f692-42b1-8d5e-512daae90cc8-images\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196323 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmkx8\" (UniqueName: \"kubernetes.io/projected/a19dcc7f-c3e9-4aa8-90dc-412550a8060f-kube-api-access-vmkx8\") pod \"migrator-59844c95c7-hjbqw\" (UID: \"a19dcc7f-c3e9-4aa8-90dc-412550a8060f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196338 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvpn8\" (UniqueName: \"kubernetes.io/projected/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-kube-api-access-mvpn8\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196355 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-config\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196370 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3651bf5e-f692-42b1-8d5e-512daae90cc8-proxy-tls\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196403 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9350aae-3053-431e-a2d6-2137f990ca08-config\") pod \"kube-apiserver-operator-766d6c64bb-9h9cg\" (UID: \"c9350aae-3053-431e-a2d6-2137f990ca08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196421 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8911791-9db1-4463-997e-1ed50da17324-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196439 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg8sc\" (UniqueName: \"kubernetes.io/projected/596a60f1-3eb7-4a57-af4a-1cb1f37f2824-kube-api-access-vg8sc\") pod \"package-server-manager-789f6589d5-ljbnb\" (UID: \"596a60f1-3eb7-4a57-af4a-1cb1f37f2824\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196456 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196474 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b189330a-ee63-45f1-8104-4ef173f8ee22-etcd-client\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196490 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-config\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196507 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/977bd2bf-e652-4b16-b8fc-902d4a1d7860-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nnjp\" (UID: \"977bd2bf-e652-4b16-b8fc-902d4a1d7860\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196531 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-serving-cert\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196548 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25db14a1-725f-42ae-a6e9-646546b584c7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196565 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32de8698-4bd5-4154-92a3-76930504a72d-serving-cert\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196581 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvj8q\" (UniqueName: \"kubernetes.io/projected/1eabf562-d289-4685-8ee5-ed1525930d19-kube-api-access-gvj8q\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196597 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5n78\" (UniqueName: \"kubernetes.io/projected/4d53a745-f985-4659-b62c-ce297ce8ce85-kube-api-access-r5n78\") pod \"service-ca-9c57cc56f-6kx9j\" (UID: \"4d53a745-f985-4659-b62c-ce297ce8ce85\") " pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196615 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196631 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b189330a-ee63-45f1-8104-4ef173f8ee22-audit-policies\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196659 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8911791-9db1-4463-997e-1ed50da17324-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196677 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196694 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4d53a745-f985-4659-b62c-ce297ce8ce85-signing-cabundle\") pod \"service-ca-9c57cc56f-6kx9j\" (UID: \"4d53a745-f985-4659-b62c-ce297ce8ce85\") " pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196710 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b8f105b-569d-47f2-b564-a0830b010e31-service-ca-bundle\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196728 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b189330a-ee63-45f1-8104-4ef173f8ee22-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196747 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196764 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr5xr\" (UniqueName: \"kubernetes.io/projected/8a35921c-db91-45c2-a92c-a42f5b2cab84-kube-api-access-gr5xr\") pod \"dns-operator-744455d44c-cdrbc\" (UID: \"8a35921c-db91-45c2-a92c-a42f5b2cab84\") " pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196788 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9d605c02-a15d-46a8-942c-cd85e6ce5452-proxy-tls\") pod \"machine-config-controller-84d6567774-nhmlj\" (UID: \"9d605c02-a15d-46a8-942c-cd85e6ce5452\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196811 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fftp\" (UniqueName: \"kubernetes.io/projected/f8911791-9db1-4463-997e-1ed50da17324-kube-api-access-6fftp\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196834 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32de8698-4bd5-4154-92a3-76930504a72d-config\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196855 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d52fce7-049e-441e-8e40-15d044e0319a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-xt6kb\" (UID: \"6d52fce7-049e-441e-8e40-15d044e0319a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196873 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b189330a-ee63-45f1-8104-4ef173f8ee22-encryption-config\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196890 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6vn7\" (UniqueName: \"kubernetes.io/projected/e6351c23-e315-4c92-a467-380da403d3c4-kube-api-access-q6vn7\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196906 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c50241f8-d135-4df8-b047-e76fb28b8a3d-srv-cert\") pod \"olm-operator-6b444d44fb-2cgk2\" (UID: \"c50241f8-d135-4df8-b047-e76fb28b8a3d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196935 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b189330a-ee63-45f1-8104-4ef173f8ee22-audit-dir\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196957 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25db14a1-725f-42ae-a6e9-646546b584c7-trusted-ca\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196975 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-etcd-client\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.196993 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-trusted-ca-bundle\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.197027 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-audit-dir\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.197050 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35efd97c-0521-428a-896d-b67490207db5-serving-cert\") pod \"openshift-config-operator-7777fb866f-7zllh\" (UID: \"35efd97c-0521-428a-896d-b67490207db5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.197073 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6351c23-e315-4c92-a467-380da403d3c4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.197097 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-serving-cert\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.197120 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-oauth-config\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.197144 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eabf562-d289-4685-8ee5-ed1525930d19-config-volume\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.197165 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c51c973-c370-41e8-b167-25d3b11418bf-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.197186 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvhcs\" (UniqueName: \"kubernetes.io/projected/6d52fce7-049e-441e-8e40-15d044e0319a-kube-api-access-nvhcs\") pod \"cluster-samples-operator-665b6dd947-xt6kb\" (UID: \"6d52fce7-049e-441e-8e40-15d044e0319a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.198034 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/00beace7-1e83-40ed-8d92-6da0cae7817f-serviceca\") pod \"image-pruner-29486880-rhzd2\" (UID: \"00beace7-1e83-40ed-8d92-6da0cae7817f\") " pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.199008 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/32de8698-4bd5-4154-92a3-76930504a72d-trusted-ca\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.199489 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-audit\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.199802 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c51c973-c370-41e8-b167-25d3b11418bf-config\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.200132 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.200566 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-config\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.201036 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8d73604-754b-4dea-9be4-3451964e5589-config\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.201068 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-client-ca\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.201141 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-dir\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.203699 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7zllh"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.204752 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-client-ca\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.205331 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32de8698-4bd5-4154-92a3-76930504a72d-config\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.208000 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-policies\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.209193 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-etcd-serving-ca\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.217228 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-oauth-serving-cert\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.222505 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-audit-dir\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.223973 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.274396 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-serving-cert\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.224895 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.225273 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b189330a-ee63-45f1-8104-4ef173f8ee22-audit-policies\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.225508 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6360940-ea9d-456d-b546-5a20af404ee5-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zc8ts\" (UID: \"a6360940-ea9d-456d-b546-5a20af404ee5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.225755 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/35efd97c-0521-428a-896d-b67490207db5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-7zllh\" (UID: \"35efd97c-0521-428a-896d-b67490207db5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.226258 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6351c23-e315-4c92-a467-380da403d3c4-service-ca-bundle\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.227199 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.230658 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.231588 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-trusted-ca-bundle\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.231606 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-encryption-config\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.231652 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b189330a-ee63-45f1-8104-4ef173f8ee22-audit-dir\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.232203 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6351c23-e315-4c92-a467-380da403d3c4-config\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.232569 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d8d73604-754b-4dea-9be4-3451964e5589-auth-proxy-config\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.233100 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6351c23-e315-4c92-a467-380da403d3c4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.240053 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d8d73604-754b-4dea-9be4-3451964e5589-machine-approver-tls\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.240099 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-node-pullsecrets\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.240476 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-config\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.243191 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b189330a-ee63-45f1-8104-4ef173f8ee22-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.244021 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-image-import-ca\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.245679 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-config\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.245977 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6360940-ea9d-456d-b546-5a20af404ee5-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zc8ts\" (UID: \"a6360940-ea9d-456d-b546-5a20af404ee5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.246604 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b189330a-ee63-45f1-8104-4ef173f8ee22-etcd-client\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.246786 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-oauth-config\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.247123 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b189330a-ee63-45f1-8104-4ef173f8ee22-encryption-config\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.247211 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37a2847-a94a-4a0c-b092-1ed7155a2d35-serving-cert\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.247233 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kvcv8"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.274881 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-serving-cert\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.248034 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32de8698-4bd5-4154-92a3-76930504a72d-serving-cert\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.274925 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jmv6d"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.248279 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.248599 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b189330a-ee63-45f1-8104-4ef173f8ee22-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.248641 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6351c23-e315-4c92-a467-380da403d3c4-serving-cert\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.275087 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c51c973-c370-41e8-b167-25d3b11418bf-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.249927 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.249556 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-serving-cert\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.266390 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9c51c973-c370-41e8-b167-25d3b11418bf-images\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.270602 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-config\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.274634 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f8911791-9db1-4463-997e-1ed50da17324-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.269122 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-etcd-client\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.275513 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.231886 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.224295 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-service-ca\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.250279 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.247521 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/35efd97c-0521-428a-896d-b67490207db5-serving-cert\") pod \"openshift-config-operator-7777fb866f-7zllh\" (UID: \"35efd97c-0521-428a-896d-b67490207db5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.248314 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b189330a-ee63-45f1-8104-4ef173f8ee22-serving-cert\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.284498 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.286913 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.288048 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2scbc"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.288092 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qzmmz"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.288118 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-lgg5l"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.288650 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.289056 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.289453 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n8shl"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.290250 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.293847 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f8911791-9db1-4463-997e-1ed50da17324-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.298536 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ft4kq"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299642 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c50241f8-d135-4df8-b047-e76fb28b8a3d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2cgk2\" (UID: \"c50241f8-d135-4df8-b047-e76fb28b8a3d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299671 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299694 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm4tl\" (UniqueName: \"kubernetes.io/projected/7af03157-ee92-4e72-a775-acaeabb73e65-kube-api-access-jm4tl\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299712 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6kdj\" (UniqueName: \"kubernetes.io/projected/9d605c02-a15d-46a8-942c-cd85e6ce5452-kube-api-access-h6kdj\") pod \"machine-config-controller-84d6567774-nhmlj\" (UID: \"9d605c02-a15d-46a8-942c-cd85e6ce5452\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299728 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4d53a745-f985-4659-b62c-ce297ce8ce85-signing-key\") pod \"service-ca-9c57cc56f-6kx9j\" (UID: \"4d53a745-f985-4659-b62c-ce297ce8ce85\") " pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299751 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299789 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2d89ff2-f919-459e-8089-5097aab0f4e2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sxjwn\" (UID: \"c2d89ff2-f919-459e-8089-5097aab0f4e2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299806 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9d605c02-a15d-46a8-942c-cd85e6ce5452-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nhmlj\" (UID: \"9d605c02-a15d-46a8-942c-cd85e6ce5452\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299822 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299836 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9350aae-3053-431e-a2d6-2137f990ca08-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9h9cg\" (UID: \"c9350aae-3053-431e-a2d6-2137f990ca08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299852 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9350aae-3053-431e-a2d6-2137f990ca08-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9h9cg\" (UID: \"c9350aae-3053-431e-a2d6-2137f990ca08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299873 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d8eed212-9137-45e5-8347-1f921fbedb19-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ck624\" (UID: \"d8eed212-9137-45e5-8347-1f921fbedb19\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299895 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm9zj\" (UniqueName: \"kubernetes.io/projected/aca1e4a5-f702-4803-8f47-7fcb8c7326b6-kube-api-access-mm9zj\") pod \"catalog-operator-68c6474976-2xb97\" (UID: \"aca1e4a5-f702-4803-8f47-7fcb8c7326b6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299912 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299930 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3651bf5e-f692-42b1-8d5e-512daae90cc8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299945 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d89ff2-f919-459e-8089-5097aab0f4e2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sxjwn\" (UID: \"c2d89ff2-f919-459e-8089-5097aab0f4e2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299961 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96br9\" (UniqueName: \"kubernetes.io/projected/25db14a1-725f-42ae-a6e9-646546b584c7-kube-api-access-96br9\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299982 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-tmpfs\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.299999 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7b8f105b-569d-47f2-b564-a0830b010e31-default-certificate\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300016 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/aca1e4a5-f702-4803-8f47-7fcb8c7326b6-profile-collector-cert\") pod \"catalog-operator-68c6474976-2xb97\" (UID: \"aca1e4a5-f702-4803-8f47-7fcb8c7326b6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300030 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gszgz\" (UniqueName: \"kubernetes.io/projected/c50241f8-d135-4df8-b047-e76fb28b8a3d-kube-api-access-gszgz\") pod \"olm-operator-6b444d44fb-2cgk2\" (UID: \"c50241f8-d135-4df8-b047-e76fb28b8a3d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300062 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300087 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eabf562-d289-4685-8ee5-ed1525930d19-secret-volume\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300104 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-webhook-cert\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300125 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7b8f105b-569d-47f2-b564-a0830b010e31-stats-auth\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300139 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcwrj\" (UniqueName: \"kubernetes.io/projected/3651bf5e-f692-42b1-8d5e-512daae90cc8-kube-api-access-dcwrj\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300155 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b8f105b-569d-47f2-b564-a0830b010e31-metrics-certs\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300174 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-config\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300197 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300213 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxkrp\" (UniqueName: \"kubernetes.io/projected/4b605a66-7904-4596-a67f-ea21ef41a24b-kube-api-access-dxkrp\") pod \"downloads-7954f5f757-6m9lm\" (UID: \"4b605a66-7904-4596-a67f-ea21ef41a24b\") " pod="openshift-console/downloads-7954f5f757-6m9lm" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300231 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/596a60f1-3eb7-4a57-af4a-1cb1f37f2824-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ljbnb\" (UID: \"596a60f1-3eb7-4a57-af4a-1cb1f37f2824\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300251 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300266 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d89ff2-f919-459e-8089-5097aab0f4e2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sxjwn\" (UID: \"c2d89ff2-f919-459e-8089-5097aab0f4e2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300282 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmkx8\" (UniqueName: \"kubernetes.io/projected/a19dcc7f-c3e9-4aa8-90dc-412550a8060f-kube-api-access-vmkx8\") pod \"migrator-59844c95c7-hjbqw\" (UID: \"a19dcc7f-c3e9-4aa8-90dc-412550a8060f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300297 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3651bf5e-f692-42b1-8d5e-512daae90cc8-images\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.300312 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3651bf5e-f692-42b1-8d5e-512daae90cc8-proxy-tls\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.301251 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.301520 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-tmpfs\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.302082 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3651bf5e-f692-42b1-8d5e-512daae90cc8-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.302762 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9d605c02-a15d-46a8-942c-cd85e6ce5452-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-nhmlj\" (UID: \"9d605c02-a15d-46a8-942c-cd85e6ce5452\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.302765 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303029 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvpn8\" (UniqueName: \"kubernetes.io/projected/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-kube-api-access-mvpn8\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303064 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9350aae-3053-431e-a2d6-2137f990ca08-config\") pod \"kube-apiserver-operator-766d6c64bb-9h9cg\" (UID: \"c9350aae-3053-431e-a2d6-2137f990ca08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303082 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg8sc\" (UniqueName: \"kubernetes.io/projected/596a60f1-3eb7-4a57-af4a-1cb1f37f2824-kube-api-access-vg8sc\") pod \"package-server-manager-789f6589d5-ljbnb\" (UID: \"596a60f1-3eb7-4a57-af4a-1cb1f37f2824\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303099 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/977bd2bf-e652-4b16-b8fc-902d4a1d7860-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nnjp\" (UID: \"977bd2bf-e652-4b16-b8fc-902d4a1d7860\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303116 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25db14a1-725f-42ae-a6e9-646546b584c7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303134 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvj8q\" (UniqueName: \"kubernetes.io/projected/1eabf562-d289-4685-8ee5-ed1525930d19-kube-api-access-gvj8q\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303153 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5n78\" (UniqueName: \"kubernetes.io/projected/4d53a745-f985-4659-b62c-ce297ce8ce85-kube-api-access-r5n78\") pod \"service-ca-9c57cc56f-6kx9j\" (UID: \"4d53a745-f985-4659-b62c-ce297ce8ce85\") " pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303173 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4d53a745-f985-4659-b62c-ce297ce8ce85-signing-cabundle\") pod \"service-ca-9c57cc56f-6kx9j\" (UID: \"4d53a745-f985-4659-b62c-ce297ce8ce85\") " pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303191 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr5xr\" (UniqueName: \"kubernetes.io/projected/8a35921c-db91-45c2-a92c-a42f5b2cab84-kube-api-access-gr5xr\") pod \"dns-operator-744455d44c-cdrbc\" (UID: \"8a35921c-db91-45c2-a92c-a42f5b2cab84\") " pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303206 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9d605c02-a15d-46a8-942c-cd85e6ce5452-proxy-tls\") pod \"machine-config-controller-84d6567774-nhmlj\" (UID: \"9d605c02-a15d-46a8-942c-cd85e6ce5452\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303221 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b8f105b-569d-47f2-b564-a0830b010e31-service-ca-bundle\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303238 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303258 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d52fce7-049e-441e-8e40-15d044e0319a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-xt6kb\" (UID: \"6d52fce7-049e-441e-8e40-15d044e0319a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303274 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c50241f8-d135-4df8-b047-e76fb28b8a3d-srv-cert\") pod \"olm-operator-6b444d44fb-2cgk2\" (UID: \"c50241f8-d135-4df8-b047-e76fb28b8a3d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303306 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25db14a1-725f-42ae-a6e9-646546b584c7-trusted-ca\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303331 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eabf562-d289-4685-8ee5-ed1525930d19-config-volume\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303345 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvhcs\" (UniqueName: \"kubernetes.io/projected/6d52fce7-049e-441e-8e40-15d044e0319a-kube-api-access-nvhcs\") pod \"cluster-samples-operator-665b6dd947-xt6kb\" (UID: \"6d52fce7-049e-441e-8e40-15d044e0319a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303362 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25db14a1-725f-42ae-a6e9-646546b584c7-metrics-tls\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303402 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk6nf\" (UniqueName: \"kubernetes.io/projected/d8eed212-9137-45e5-8347-1f921fbedb19-kube-api-access-nk6nf\") pod \"multus-admission-controller-857f4d67dd-ck624\" (UID: \"d8eed212-9137-45e5-8347-1f921fbedb19\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303421 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlm94\" (UniqueName: \"kubernetes.io/projected/89435826-645d-48a2-aa3b-f5c42003dcbe-kube-api-access-tlm94\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303448 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/aca1e4a5-f702-4803-8f47-7fcb8c7326b6-srv-cert\") pod \"catalog-operator-68c6474976-2xb97\" (UID: \"aca1e4a5-f702-4803-8f47-7fcb8c7326b6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303463 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977bd2bf-e652-4b16-b8fc-902d4a1d7860-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nnjp\" (UID: \"977bd2bf-e652-4b16-b8fc-902d4a1d7860\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303489 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303504 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7srgg\" (UniqueName: \"kubernetes.io/projected/977bd2bf-e652-4b16-b8fc-902d4a1d7860-kube-api-access-7srgg\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nnjp\" (UID: \"977bd2bf-e652-4b16-b8fc-902d4a1d7860\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303521 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-apiservice-cert\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303541 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a35921c-db91-45c2-a92c-a42f5b2cab84-metrics-tls\") pod \"dns-operator-744455d44c-cdrbc\" (UID: \"8a35921c-db91-45c2-a92c-a42f5b2cab84\") " pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303558 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4bc9\" (UniqueName: \"kubernetes.io/projected/7b8f105b-569d-47f2-b564-a0830b010e31-kube-api-access-j4bc9\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.303117 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.304170 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3651bf5e-f692-42b1-8d5e-512daae90cc8-images\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.304848 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29486880-rhzd2"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.306564 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.308840 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8a35921c-db91-45c2-a92c-a42f5b2cab84-metrics-tls\") pod \"dns-operator-744455d44c-cdrbc\" (UID: \"8a35921c-db91-45c2-a92c-a42f5b2cab84\") " pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.309248 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3651bf5e-f692-42b1-8d5e-512daae90cc8-proxy-tls\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.310054 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.310187 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.313165 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.314277 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cdrbc"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.316635 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.317543 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.317731 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-spx5f"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.319113 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-spx5f" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.319332 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ck624"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.323358 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-g2smk"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.325820 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.325991 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6m9lm"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.326215 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.327272 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/aca1e4a5-f702-4803-8f47-7fcb8c7326b6-srv-cert\") pod \"catalog-operator-68c6474976-2xb97\" (UID: \"aca1e4a5-f702-4803-8f47-7fcb8c7326b6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.327799 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.328342 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.329732 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6kx9j"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.330678 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.331651 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wvvpg"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.332938 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.334079 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-x649d"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.335504 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.336468 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.337611 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.338586 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.339633 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.340669 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.341698 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-w7sr4"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.342670 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.343140 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rsw66"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.344192 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.345172 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-64phz"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.346281 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.346370 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.346512 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.347453 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-spx5f"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.348494 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.349452 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.350522 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.351874 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-64phz"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.352785 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.353753 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n8shl"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.354774 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-w7sr4"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.355874 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shf2w"] Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.366912 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.375406 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/c50241f8-d135-4df8-b047-e76fb28b8a3d-profile-collector-cert\") pod \"olm-operator-6b444d44fb-2cgk2\" (UID: \"c50241f8-d135-4df8-b047-e76fb28b8a3d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.375456 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eabf562-d289-4685-8ee5-ed1525930d19-secret-volume\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.376106 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/aca1e4a5-f702-4803-8f47-7fcb8c7326b6-profile-collector-cert\") pod \"catalog-operator-68c6474976-2xb97\" (UID: \"aca1e4a5-f702-4803-8f47-7fcb8c7326b6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.387089 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.406659 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.426751 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.439545 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6d52fce7-049e-441e-8e40-15d044e0319a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-xt6kb\" (UID: \"6d52fce7-049e-441e-8e40-15d044e0319a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.447505 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.467772 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.486314 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.498615 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25db14a1-725f-42ae-a6e9-646546b584c7-metrics-tls\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.506943 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.527572 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.546466 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.572294 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.576681 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/25db14a1-725f-42ae-a6e9-646546b584c7-trusted-ca\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.586462 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.594994 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d8eed212-9137-45e5-8347-1f921fbedb19-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ck624\" (UID: \"d8eed212-9137-45e5-8347-1f921fbedb19\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.606227 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.626632 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.642026 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/596a60f1-3eb7-4a57-af4a-1cb1f37f2824-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-ljbnb\" (UID: \"596a60f1-3eb7-4a57-af4a-1cb1f37f2824\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.647209 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.667574 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.687592 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.707799 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.727503 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.735627 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7b8f105b-569d-47f2-b564-a0830b010e31-default-certificate\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.748619 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.755109 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7b8f105b-569d-47f2-b564-a0830b010e31-stats-auth\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.767877 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.777437 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b8f105b-569d-47f2-b564-a0830b010e31-metrics-certs\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.787207 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.795923 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b8f105b-569d-47f2-b564-a0830b010e31-service-ca-bundle\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.808142 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.847797 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.868455 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.886804 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.895277 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2d89ff2-f919-459e-8089-5097aab0f4e2-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sxjwn\" (UID: \"c2d89ff2-f919-459e-8089-5097aab0f4e2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.908045 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.915557 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2d89ff2-f919-459e-8089-5097aab0f4e2-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sxjwn\" (UID: \"c2d89ff2-f919-459e-8089-5097aab0f4e2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.927037 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.935934 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-webhook-cert\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.938632 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-apiservice-cert\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.947431 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.967484 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.977297 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4d53a745-f985-4659-b62c-ce297ce8ce85-signing-key\") pod \"service-ca-9c57cc56f-6kx9j\" (UID: \"4d53a745-f985-4659-b62c-ce297ce8ce85\") " pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:50 crc kubenswrapper[4676]: I0124 00:05:50.987029 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.007872 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.027933 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.035685 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4d53a745-f985-4659-b62c-ce297ce8ce85-signing-cabundle\") pod \"service-ca-9c57cc56f-6kx9j\" (UID: \"4d53a745-f985-4659-b62c-ce297ce8ce85\") " pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.047201 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.058855 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9d605c02-a15d-46a8-942c-cd85e6ce5452-proxy-tls\") pod \"machine-config-controller-84d6567774-nhmlj\" (UID: \"9d605c02-a15d-46a8-942c-cd85e6ce5452\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.067199 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.087407 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.099342 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/c50241f8-d135-4df8-b047-e76fb28b8a3d-srv-cert\") pod \"olm-operator-6b444d44fb-2cgk2\" (UID: \"c50241f8-d135-4df8-b047-e76fb28b8a3d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.106916 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.116761 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/977bd2bf-e652-4b16-b8fc-902d4a1d7860-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nnjp\" (UID: \"977bd2bf-e652-4b16-b8fc-902d4a1d7860\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.127691 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.137617 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/977bd2bf-e652-4b16-b8fc-902d4a1d7860-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nnjp\" (UID: \"977bd2bf-e652-4b16-b8fc-902d4a1d7860\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.148323 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.165833 4676 request.go:700] Waited for 1.016067431s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.167303 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.186518 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.206890 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.233715 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.247441 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.255998 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c9350aae-3053-431e-a2d6-2137f990ca08-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-9h9cg\" (UID: \"c9350aae-3053-431e-a2d6-2137f990ca08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.269875 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.274471 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9350aae-3053-431e-a2d6-2137f990ca08-config\") pod \"kube-apiserver-operator-766d6c64bb-9h9cg\" (UID: \"c9350aae-3053-431e-a2d6-2137f990ca08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.287502 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.302265 4676 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.302289 4676 secret.go:188] Couldn't get secret openshift-kube-controller-manager-operator/kube-controller-manager-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.302352 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-trusted-ca podName:7af03157-ee92-4e72-a775-acaeabb73e65 nodeName:}" failed. No retries permitted until 2026-01-24 00:05:51.802325304 +0000 UTC m=+135.832296345 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-trusted-ca") pod "marketplace-operator-79b997595-rsw66" (UID: "7af03157-ee92-4e72-a775-acaeabb73e65") : failed to sync configmap cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.302356 4676 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.302414 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-serving-cert podName:2f5e70d9-8b16-4684-bd98-4287ccbb6d85 nodeName:}" failed. No retries permitted until 2026-01-24 00:05:51.802367135 +0000 UTC m=+135.832338176 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-serving-cert") pod "kube-controller-manager-operator-78b949d7b-975pl" (UID: "2f5e70d9-8b16-4684-bd98-4287ccbb6d85") : failed to sync secret cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.302452 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-operator-metrics podName:7af03157-ee92-4e72-a775-acaeabb73e65 nodeName:}" failed. No retries permitted until 2026-01-24 00:05:51.802433797 +0000 UTC m=+135.832404838 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-operator-metrics") pod "marketplace-operator-79b997595-rsw66" (UID: "7af03157-ee92-4e72-a775-acaeabb73e65") : failed to sync secret cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.303656 4676 configmap.go:193] Couldn't get configMap openshift-kube-controller-manager-operator/kube-controller-manager-operator-config: failed to sync configmap cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.303729 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-config podName:2f5e70d9-8b16-4684-bd98-4287ccbb6d85 nodeName:}" failed. No retries permitted until 2026-01-24 00:05:51.803704589 +0000 UTC m=+135.833675640 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-config") pod "kube-controller-manager-operator-78b949d7b-975pl" (UID: "2f5e70d9-8b16-4684-bd98-4287ccbb6d85") : failed to sync configmap cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.305707 4676 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: E0124 00:05:51.305944 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1eabf562-d289-4685-8ee5-ed1525930d19-config-volume podName:1eabf562-d289-4685-8ee5-ed1525930d19 nodeName:}" failed. No retries permitted until 2026-01-24 00:05:51.80593158 +0000 UTC m=+135.835902591 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/1eabf562-d289-4685-8ee5-ed1525930d19-config-volume") pod "collect-profiles-29486880-7srrg" (UID: "1eabf562-d289-4685-8ee5-ed1525930d19") : failed to sync configmap cache: timed out waiting for the condition Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.307578 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.327303 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.347088 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.377870 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.387775 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.407052 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.427603 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.447198 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.467095 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.487612 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.507672 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.527077 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.547690 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.567013 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.586607 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.606774 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.626616 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.647073 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.668671 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.687558 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.707718 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.767146 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv9mz\" (UniqueName: \"kubernetes.io/projected/b37a2847-a94a-4a0c-b092-1ed7155a2d35-kube-api-access-nv9mz\") pod \"controller-manager-879f6c89f-qzmmz\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.787591 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hzh\" (UniqueName: \"kubernetes.io/projected/9c51c973-c370-41e8-b167-25d3b11418bf-kube-api-access-87hzh\") pod \"machine-api-operator-5694c8668f-jmv6d\" (UID: \"9c51c973-c370-41e8-b167-25d3b11418bf\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.801661 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.804636 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f9mm\" (UniqueName: \"kubernetes.io/projected/2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e-kube-api-access-5f9mm\") pod \"apiserver-76f77b778f-2scbc\" (UID: \"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e\") " pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.818675 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.822067 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmt92\" (UniqueName: \"kubernetes.io/projected/d8d73604-754b-4dea-9be4-3451964e5589-kube-api-access-cmt92\") pod \"machine-approver-56656f9798-ml6vn\" (UID: \"d8d73604-754b-4dea-9be4-3451964e5589\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.833989 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eabf562-d289-4685-8ee5-ed1525930d19-config-volume\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.834260 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.834417 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.834541 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.834596 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-config\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.834731 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eabf562-d289-4685-8ee5-ed1525930d19-config-volume\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.835597 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-config\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.837679 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.838345 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.840040 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.845444 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th8cd\" (UniqueName: \"kubernetes.io/projected/b189330a-ee63-45f1-8104-4ef173f8ee22-kube-api-access-th8cd\") pod \"apiserver-7bbb656c7d-7d9xm\" (UID: \"b189330a-ee63-45f1-8104-4ef173f8ee22\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.867352 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.867765 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p6c5\" (UniqueName: \"kubernetes.io/projected/35efd97c-0521-428a-896d-b67490207db5-kube-api-access-6p6c5\") pod \"openshift-config-operator-7777fb866f-7zllh\" (UID: \"35efd97c-0521-428a-896d-b67490207db5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.883414 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qp87\" (UniqueName: \"kubernetes.io/projected/00beace7-1e83-40ed-8d92-6da0cae7817f-kube-api-access-4qp87\") pod \"image-pruner-29486880-rhzd2\" (UID: \"00beace7-1e83-40ed-8d92-6da0cae7817f\") " pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.884975 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.910640 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ksdz\" (UniqueName: \"kubernetes.io/projected/32de8698-4bd5-4154-92a3-76930504a72d-kube-api-access-4ksdz\") pod \"console-operator-58897d9998-x649d\" (UID: \"32de8698-4bd5-4154-92a3-76930504a72d\") " pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.921916 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.929198 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6dpw\" (UniqueName: \"kubernetes.io/projected/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-kube-api-access-f6dpw\") pod \"console-f9d7485db-g2smk\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.947979 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.949661 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4cw9\" (UniqueName: \"kubernetes.io/projected/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-kube-api-access-h4cw9\") pod \"route-controller-manager-6576b87f9c-n6vx5\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.962741 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6vn7\" (UniqueName: \"kubernetes.io/projected/e6351c23-e315-4c92-a467-380da403d3c4-kube-api-access-q6vn7\") pod \"authentication-operator-69f744f599-ft4kq\" (UID: \"e6351c23-e315-4c92-a467-380da403d3c4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.986864 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqnsp\" (UniqueName: \"kubernetes.io/projected/a6360940-ea9d-456d-b546-5a20af404ee5-kube-api-access-jqnsp\") pod \"openshift-apiserver-operator-796bbdcf4f-zc8ts\" (UID: \"a6360940-ea9d-456d-b546-5a20af404ee5\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:51 crc kubenswrapper[4676]: I0124 00:05:51.994793 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.004310 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fftp\" (UniqueName: \"kubernetes.io/projected/f8911791-9db1-4463-997e-1ed50da17324-kube-api-access-6fftp\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.028860 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.028665 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.032586 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f8911791-9db1-4463-997e-1ed50da17324-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rkpbj\" (UID: \"f8911791-9db1-4463-997e-1ed50da17324\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.043025 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-jmv6d"] Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.047858 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.071229 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.084891 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.088577 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.091790 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.101709 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.106888 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.116725 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qzmmz"] Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.119931 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.129967 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.144845 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2scbc"] Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.147081 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.164330 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" event={"ID":"d8d73604-754b-4dea-9be4-3451964e5589","Type":"ContainerStarted","Data":"14117bd130e68b0fc47a586d86a9175abdcd7dcd9d55c2e32081a4637f074f9e"} Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.170921 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.187253 4676 request.go:700] Waited for 1.885694613s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/serviceaccounts/olm-operator-serviceaccount/token Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.190070 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" event={"ID":"b37a2847-a94a-4a0c-b092-1ed7155a2d35","Type":"ContainerStarted","Data":"63dfdd35be2732ab824c6cbb0405da330ff5fefbf093c8f9f68db2d40a451da2"} Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.211064 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm9zj\" (UniqueName: \"kubernetes.io/projected/aca1e4a5-f702-4803-8f47-7fcb8c7326b6-kube-api-access-mm9zj\") pod \"catalog-operator-68c6474976-2xb97\" (UID: \"aca1e4a5-f702-4803-8f47-7fcb8c7326b6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.212869 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" event={"ID":"9c51c973-c370-41e8-b167-25d3b11418bf","Type":"ContainerStarted","Data":"9dea6e623cd16e538adb931ab7cfe35f12d462deb041c46eff49d5d7b18c7cde"} Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.217186 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-7zllh"] Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.224702 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c9350aae-3053-431e-a2d6-2137f990ca08-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-9h9cg\" (UID: \"c9350aae-3053-431e-a2d6-2137f990ca08\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.244158 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gszgz\" (UniqueName: \"kubernetes.io/projected/c50241f8-d135-4df8-b047-e76fb28b8a3d-kube-api-access-gszgz\") pod \"olm-operator-6b444d44fb-2cgk2\" (UID: \"c50241f8-d135-4df8-b047-e76fb28b8a3d\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.253016 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.277802 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.293902 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcwrj\" (UniqueName: \"kubernetes.io/projected/3651bf5e-f692-42b1-8d5e-512daae90cc8-kube-api-access-dcwrj\") pod \"machine-config-operator-74547568cd-6wjpz\" (UID: \"3651bf5e-f692-42b1-8d5e-512daae90cc8\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.302984 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96br9\" (UniqueName: \"kubernetes.io/projected/25db14a1-725f-42ae-a6e9-646546b584c7-kube-api-access-96br9\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.309447 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6kdj\" (UniqueName: \"kubernetes.io/projected/9d605c02-a15d-46a8-942c-cd85e6ce5452-kube-api-access-h6kdj\") pod \"machine-config-controller-84d6567774-nhmlj\" (UID: \"9d605c02-a15d-46a8-942c-cd85e6ce5452\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.326500 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm4tl\" (UniqueName: \"kubernetes.io/projected/7af03157-ee92-4e72-a775-acaeabb73e65-kube-api-access-jm4tl\") pod \"marketplace-operator-79b997595-rsw66\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.345538 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxkrp\" (UniqueName: \"kubernetes.io/projected/4b605a66-7904-4596-a67f-ea21ef41a24b-kube-api-access-dxkrp\") pod \"downloads-7954f5f757-6m9lm\" (UID: \"4b605a66-7904-4596-a67f-ea21ef41a24b\") " pod="openshift-console/downloads-7954f5f757-6m9lm" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.367321 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29486880-rhzd2"] Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.370103 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmkx8\" (UniqueName: \"kubernetes.io/projected/a19dcc7f-c3e9-4aa8-90dc-412550a8060f-kube-api-access-vmkx8\") pod \"migrator-59844c95c7-hjbqw\" (UID: \"a19dcc7f-c3e9-4aa8-90dc-412550a8060f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.387935 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2d89ff2-f919-459e-8089-5097aab0f4e2-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-sxjwn\" (UID: \"c2d89ff2-f919-459e-8089-5097aab0f4e2\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.413159 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4bc9\" (UniqueName: \"kubernetes.io/projected/7b8f105b-569d-47f2-b564-a0830b010e31-kube-api-access-j4bc9\") pod \"router-default-5444994796-zq745\" (UID: \"7b8f105b-569d-47f2-b564-a0830b010e31\") " pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.425461 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6m9lm" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.428232 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-x649d"] Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.436836 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvpn8\" (UniqueName: \"kubernetes.io/projected/fdb046fc-eba9-4f07-a1d1-2db71a2d46c1-kube-api-access-mvpn8\") pod \"packageserver-d55dfcdfc-r6r59\" (UID: \"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.439912 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.461239 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.463202 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg8sc\" (UniqueName: \"kubernetes.io/projected/596a60f1-3eb7-4a57-af4a-1cb1f37f2824-kube-api-access-vg8sc\") pod \"package-server-manager-789f6589d5-ljbnb\" (UID: \"596a60f1-3eb7-4a57-af4a-1cb1f37f2824\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.481829 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/25db14a1-725f-42ae-a6e9-646546b584c7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wh9bw\" (UID: \"25db14a1-725f-42ae-a6e9-646546b584c7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.492285 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvj8q\" (UniqueName: \"kubernetes.io/projected/1eabf562-d289-4685-8ee5-ed1525930d19-kube-api-access-gvj8q\") pod \"collect-profiles-29486880-7srrg\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.492653 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.501240 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.504734 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5n78\" (UniqueName: \"kubernetes.io/projected/4d53a745-f985-4659-b62c-ce297ce8ce85-kube-api-access-r5n78\") pod \"service-ca-9c57cc56f-6kx9j\" (UID: \"4d53a745-f985-4659-b62c-ce297ce8ce85\") " pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.507268 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.515967 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.524504 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.531760 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.532016 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr5xr\" (UniqueName: \"kubernetes.io/projected/8a35921c-db91-45c2-a92c-a42f5b2cab84-kube-api-access-gr5xr\") pod \"dns-operator-744455d44c-cdrbc\" (UID: \"8a35921c-db91-45c2-a92c-a42f5b2cab84\") " pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.538273 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.549151 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk6nf\" (UniqueName: \"kubernetes.io/projected/d8eed212-9137-45e5-8347-1f921fbedb19-kube-api-access-nk6nf\") pod \"multus-admission-controller-857f4d67dd-ck624\" (UID: \"d8eed212-9137-45e5-8347-1f921fbedb19\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.563580 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvhcs\" (UniqueName: \"kubernetes.io/projected/6d52fce7-049e-441e-8e40-15d044e0319a-kube-api-access-nvhcs\") pod \"cluster-samples-operator-665b6dd947-xt6kb\" (UID: \"6d52fce7-049e-441e-8e40-15d044e0319a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.570028 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.577371 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.581570 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f5e70d9-8b16-4684-bd98-4287ccbb6d85-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-975pl\" (UID: \"2f5e70d9-8b16-4684-bd98-4287ccbb6d85\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.600690 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlm94\" (UniqueName: \"kubernetes.io/projected/89435826-645d-48a2-aa3b-f5c42003dcbe-kube-api-access-tlm94\") pod \"oauth-openshift-558db77b4-kvcv8\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.603464 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.619103 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-g2smk"] Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.621173 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7srgg\" (UniqueName: \"kubernetes.io/projected/977bd2bf-e652-4b16-b8fc-902d4a1d7860-kube-api-access-7srgg\") pod \"kube-storage-version-migrator-operator-b67b599dd-4nnjp\" (UID: \"977bd2bf-e652-4b16-b8fc-902d4a1d7860\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.627506 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.647174 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.667509 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.687305 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.706859 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.727201 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.732873 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.747266 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.748306 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.786783 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.787298 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.791317 4676 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.791764 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.807893 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.821080 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm"] Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.825567 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5"] Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.846648 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.927943 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.931578 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.936465 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.936537 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-tls\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.936615 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9887557b-81eb-4651-8da2-fd34d7b0be97-ca-trust-extracted\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.936650 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-bound-sa-token\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.936829 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-certificates\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.936905 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-trusted-ca\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.936975 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9887557b-81eb-4651-8da2-fd34d7b0be97-installation-pull-secrets\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:52 crc kubenswrapper[4676]: I0124 00:05:52.937006 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlknz\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-kube-api-access-hlknz\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:52 crc kubenswrapper[4676]: E0124 00:05:52.941274 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:53.441250807 +0000 UTC m=+137.471222048 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.039835 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.040099 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:53.540069631 +0000 UTC m=+137.570040672 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040408 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae793283-426a-4e02-b96b-89c3f16f2d16-config\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040459 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-bound-sa-token\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040481 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5c752b5-1413-4c4e-aea6-5302a7c69467-cert\") pod \"ingress-canary-spx5f\" (UID: \"f5c752b5-1413-4c4e-aea6-5302a7c69467\") " pod="openshift-ingress-canary/ingress-canary-spx5f" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040502 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8ljj\" (UID: \"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040535 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fv6m\" (UniqueName: \"kubernetes.io/projected/f53afe6a-307c-4b0d-88cb-596703f35f8a-kube-api-access-4fv6m\") pod \"control-plane-machine-set-operator-78cbb6b69f-mkwj8\" (UID: \"f53afe6a-307c-4b0d-88cb-596703f35f8a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040562 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl8ks\" (UniqueName: \"kubernetes.io/projected/3e6f77cf-0bf1-4e28-a894-1ca5ee320c58-kube-api-access-cl8ks\") pod \"machine-config-server-lgg5l\" (UID: \"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58\") " pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040589 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6h9j\" (UniqueName: \"kubernetes.io/projected/e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106-kube-api-access-l6h9j\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8ljj\" (UID: \"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040606 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3e6f77cf-0bf1-4e28-a894-1ca5ee320c58-certs\") pod \"machine-config-server-lgg5l\" (UID: \"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58\") " pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040635 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f53afe6a-307c-4b0d-88cb-596703f35f8a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mkwj8\" (UID: \"f53afe6a-307c-4b0d-88cb-596703f35f8a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040656 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8ljj\" (UID: \"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040695 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ae793283-426a-4e02-b96b-89c3f16f2d16-etcd-ca\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040734 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae793283-426a-4e02-b96b-89c3f16f2d16-etcd-client\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040767 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-certificates\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040807 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-trusted-ca\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040836 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9887557b-81eb-4651-8da2-fd34d7b0be97-installation-pull-secrets\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040873 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlknz\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-kube-api-access-hlknz\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040918 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27tsj\" (UniqueName: \"kubernetes.io/projected/f5c752b5-1413-4c4e-aea6-5302a7c69467-kube-api-access-27tsj\") pod \"ingress-canary-spx5f\" (UID: \"f5c752b5-1413-4c4e-aea6-5302a7c69467\") " pod="openshift-ingress-canary/ingress-canary-spx5f" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.040996 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-959hq\" (UniqueName: \"kubernetes.io/projected/ae793283-426a-4e02-b96b-89c3f16f2d16-kube-api-access-959hq\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.041032 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.041055 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae793283-426a-4e02-b96b-89c3f16f2d16-serving-cert\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.041070 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae793283-426a-4e02-b96b-89c3f16f2d16-etcd-service-ca\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.041091 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-tls\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.041109 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3e6f77cf-0bf1-4e28-a894-1ca5ee320c58-node-bootstrap-token\") pod \"machine-config-server-lgg5l\" (UID: \"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58\") " pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.041144 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9887557b-81eb-4651-8da2-fd34d7b0be97-ca-trust-extracted\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.041572 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9887557b-81eb-4651-8da2-fd34d7b0be97-ca-trust-extracted\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.045294 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-trusted-ca\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.046144 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:53.546122725 +0000 UTC m=+137.576093926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.054343 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-tls\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.060515 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ft4kq"] Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.063229 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-certificates\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.065679 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9887557b-81eb-4651-8da2-fd34d7b0be97-installation-pull-secrets\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.070276 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj"] Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.071170 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-bound-sa-token\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.082315 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlknz\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-kube-api-access-hlknz\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.096334 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg"] Jan 24 00:05:53 crc kubenswrapper[4676]: W0124 00:05:53.127896 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6351c23_e315_4c92_a467_380da403d3c4.slice/crio-c3711f3be897a53e7f5ed62f4dfea1182824530d785a6c28ea3f9882ad5f65da WatchSource:0}: Error finding container c3711f3be897a53e7f5ed62f4dfea1182824530d785a6c28ea3f9882ad5f65da: Status 404 returned error can't find the container with id c3711f3be897a53e7f5ed62f4dfea1182824530d785a6c28ea3f9882ad5f65da Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142100 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142348 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5c752b5-1413-4c4e-aea6-5302a7c69467-cert\") pod \"ingress-canary-spx5f\" (UID: \"f5c752b5-1413-4c4e-aea6-5302a7c69467\") " pod="openshift-ingress-canary/ingress-canary-spx5f" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142441 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8ljj\" (UID: \"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142544 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fv6m\" (UniqueName: \"kubernetes.io/projected/f53afe6a-307c-4b0d-88cb-596703f35f8a-kube-api-access-4fv6m\") pod \"control-plane-machine-set-operator-78cbb6b69f-mkwj8\" (UID: \"f53afe6a-307c-4b0d-88cb-596703f35f8a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142600 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl8ks\" (UniqueName: \"kubernetes.io/projected/3e6f77cf-0bf1-4e28-a894-1ca5ee320c58-kube-api-access-cl8ks\") pod \"machine-config-server-lgg5l\" (UID: \"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58\") " pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142634 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6h9j\" (UniqueName: \"kubernetes.io/projected/e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106-kube-api-access-l6h9j\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8ljj\" (UID: \"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142664 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-registration-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142692 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3e6f77cf-0bf1-4e28-a894-1ca5ee320c58-certs\") pod \"machine-config-server-lgg5l\" (UID: \"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58\") " pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142725 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f53afe6a-307c-4b0d-88cb-596703f35f8a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mkwj8\" (UID: \"f53afe6a-307c-4b0d-88cb-596703f35f8a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142752 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c-config\") pod \"service-ca-operator-777779d784-n8shl\" (UID: \"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142797 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-mountpoint-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142821 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8ljj\" (UID: \"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142853 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ae793283-426a-4e02-b96b-89c3f16f2d16-etcd-ca\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.142942 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae793283-426a-4e02-b96b-89c3f16f2d16-etcd-client\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.143033 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:53.642992486 +0000 UTC m=+137.672963487 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143364 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs2tp\" (UniqueName: \"kubernetes.io/projected/290fd523-8c24-458a-8abb-e32ca43caae1-kube-api-access-qs2tp\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143441 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce7f5025-ebd8-4cbf-af5a-460fbeb681d1-config-volume\") pod \"dns-default-w7sr4\" (UID: \"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1\") " pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143480 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-socket-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143510 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27tsj\" (UniqueName: \"kubernetes.io/projected/f5c752b5-1413-4c4e-aea6-5302a7c69467-kube-api-access-27tsj\") pod \"ingress-canary-spx5f\" (UID: \"f5c752b5-1413-4c4e-aea6-5302a7c69467\") " pod="openshift-ingress-canary/ingress-canary-spx5f" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143615 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-csi-data-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143652 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-959hq\" (UniqueName: \"kubernetes.io/projected/ae793283-426a-4e02-b96b-89c3f16f2d16-kube-api-access-959hq\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143683 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143708 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpntf\" (UniqueName: \"kubernetes.io/projected/ce7f5025-ebd8-4cbf-af5a-460fbeb681d1-kube-api-access-xpntf\") pod \"dns-default-w7sr4\" (UID: \"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1\") " pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143732 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-plugins-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143753 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c-serving-cert\") pod \"service-ca-operator-777779d784-n8shl\" (UID: \"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143777 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g949x\" (UniqueName: \"kubernetes.io/projected/bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c-kube-api-access-g949x\") pod \"service-ca-operator-777779d784-n8shl\" (UID: \"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143814 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae793283-426a-4e02-b96b-89c3f16f2d16-serving-cert\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143834 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae793283-426a-4e02-b96b-89c3f16f2d16-etcd-service-ca\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143871 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3e6f77cf-0bf1-4e28-a894-1ca5ee320c58-node-bootstrap-token\") pod \"machine-config-server-lgg5l\" (UID: \"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58\") " pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.143978 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ce7f5025-ebd8-4cbf-af5a-460fbeb681d1-metrics-tls\") pod \"dns-default-w7sr4\" (UID: \"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1\") " pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.144034 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae793283-426a-4e02-b96b-89c3f16f2d16-config\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.144677 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae793283-426a-4e02-b96b-89c3f16f2d16-config\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.146495 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8ljj\" (UID: \"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.152774 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f53afe6a-307c-4b0d-88cb-596703f35f8a-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-mkwj8\" (UID: \"f53afe6a-307c-4b0d-88cb-596703f35f8a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.153398 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ae793283-426a-4e02-b96b-89c3f16f2d16-etcd-service-ca\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.155159 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:53.655140856 +0000 UTC m=+137.685111857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.159311 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8ljj\" (UID: \"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.159416 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae793283-426a-4e02-b96b-89c3f16f2d16-etcd-client\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.164462 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3e6f77cf-0bf1-4e28-a894-1ca5ee320c58-certs\") pod \"machine-config-server-lgg5l\" (UID: \"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58\") " pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.167323 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae793283-426a-4e02-b96b-89c3f16f2d16-serving-cert\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.168811 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ae793283-426a-4e02-b96b-89c3f16f2d16-etcd-ca\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.168856 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5c752b5-1413-4c4e-aea6-5302a7c69467-cert\") pod \"ingress-canary-spx5f\" (UID: \"f5c752b5-1413-4c4e-aea6-5302a7c69467\") " pod="openshift-ingress-canary/ingress-canary-spx5f" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.194915 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fv6m\" (UniqueName: \"kubernetes.io/projected/f53afe6a-307c-4b0d-88cb-596703f35f8a-kube-api-access-4fv6m\") pod \"control-plane-machine-set-operator-78cbb6b69f-mkwj8\" (UID: \"f53afe6a-307c-4b0d-88cb-596703f35f8a\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.201619 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3e6f77cf-0bf1-4e28-a894-1ca5ee320c58-node-bootstrap-token\") pod \"machine-config-server-lgg5l\" (UID: \"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58\") " pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.207627 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl8ks\" (UniqueName: \"kubernetes.io/projected/3e6f77cf-0bf1-4e28-a894-1ca5ee320c58-kube-api-access-cl8ks\") pod \"machine-config-server-lgg5l\" (UID: \"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58\") " pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.214042 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lgg5l" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.246490 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6h9j\" (UniqueName: \"kubernetes.io/projected/e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106-kube-api-access-l6h9j\") pod \"openshift-controller-manager-operator-756b6f6bc6-q8ljj\" (UID: \"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.246661 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.246875 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs2tp\" (UniqueName: \"kubernetes.io/projected/290fd523-8c24-458a-8abb-e32ca43caae1-kube-api-access-qs2tp\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.246901 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce7f5025-ebd8-4cbf-af5a-460fbeb681d1-config-volume\") pod \"dns-default-w7sr4\" (UID: \"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1\") " pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.246919 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-socket-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.246954 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-csi-data-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.246996 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:53.746968816 +0000 UTC m=+137.776939817 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247037 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247052 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-csi-data-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247071 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpntf\" (UniqueName: \"kubernetes.io/projected/ce7f5025-ebd8-4cbf-af5a-460fbeb681d1-kube-api-access-xpntf\") pod \"dns-default-w7sr4\" (UID: \"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1\") " pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247109 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-plugins-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247140 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c-serving-cert\") pod \"service-ca-operator-777779d784-n8shl\" (UID: \"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247162 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g949x\" (UniqueName: \"kubernetes.io/projected/bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c-kube-api-access-g949x\") pod \"service-ca-operator-777779d784-n8shl\" (UID: \"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247238 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ce7f5025-ebd8-4cbf-af5a-460fbeb681d1-metrics-tls\") pod \"dns-default-w7sr4\" (UID: \"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1\") " pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247319 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-registration-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247518 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c-config\") pod \"service-ca-operator-777779d784-n8shl\" (UID: \"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247546 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-mountpoint-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.247712 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-mountpoint-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.248023 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:53.748015728 +0000 UTC m=+137.777986729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.248608 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-plugins-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.249180 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-socket-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.249244 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/290fd523-8c24-458a-8abb-e32ca43caae1-registration-dir\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.249537 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c-config\") pod \"service-ca-operator-777779d784-n8shl\" (UID: \"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.250032 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce7f5025-ebd8-4cbf-af5a-460fbeb681d1-config-volume\") pod \"dns-default-w7sr4\" (UID: \"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1\") " pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.255816 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c-serving-cert\") pod \"service-ca-operator-777779d784-n8shl\" (UID: \"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.271996 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27tsj\" (UniqueName: \"kubernetes.io/projected/f5c752b5-1413-4c4e-aea6-5302a7c69467-kube-api-access-27tsj\") pod \"ingress-canary-spx5f\" (UID: \"f5c752b5-1413-4c4e-aea6-5302a7c69467\") " pod="openshift-ingress-canary/ingress-canary-spx5f" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.272536 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ce7f5025-ebd8-4cbf-af5a-460fbeb681d1-metrics-tls\") pod \"dns-default-w7sr4\" (UID: \"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1\") " pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.304768 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs2tp\" (UniqueName: \"kubernetes.io/projected/290fd523-8c24-458a-8abb-e32ca43caae1-kube-api-access-qs2tp\") pod \"csi-hostpathplugin-64phz\" (UID: \"290fd523-8c24-458a-8abb-e32ca43caae1\") " pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.331703 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-959hq\" (UniqueName: \"kubernetes.io/projected/ae793283-426a-4e02-b96b-89c3f16f2d16-kube-api-access-959hq\") pod \"etcd-operator-b45778765-wvvpg\" (UID: \"ae793283-426a-4e02-b96b-89c3f16f2d16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.333307 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" event={"ID":"9c51c973-c370-41e8-b167-25d3b11418bf","Type":"ContainerStarted","Data":"6a43e179640e1b1f56bcefa6613ceeea5c74f898c5a73adf794aa37a8b6aca24"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.334504 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" event={"ID":"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93","Type":"ContainerStarted","Data":"caafeb67c7aefa20810226a8341333cb55d2bc2743763b95b6989d1d99cdf707"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.343623 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" event={"ID":"b37a2847-a94a-4a0c-b092-1ed7155a2d35","Type":"ContainerStarted","Data":"75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.347254 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g949x\" (UniqueName: \"kubernetes.io/projected/bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c-kube-api-access-g949x\") pod \"service-ca-operator-777779d784-n8shl\" (UID: \"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.347680 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.348162 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.348442 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:53.848430874 +0000 UTC m=+137.878401875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.348922 4676 generic.go:334] "Generic (PLEG): container finished" podID="2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e" containerID="9dbbda8f061aa866b6e59cc247c2b8e94fec1c21500f03dbd015022a26e987fe" exitCode=0 Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.348971 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" event={"ID":"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e","Type":"ContainerDied","Data":"9dbbda8f061aa866b6e59cc247c2b8e94fec1c21500f03dbd015022a26e987fe"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.348988 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" event={"ID":"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e","Type":"ContainerStarted","Data":"3c4296fbfe84777bd64c3d2c2c56a32e0594f1f75477226e668665f31814a7fc"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.351679 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-x649d" event={"ID":"32de8698-4bd5-4154-92a3-76930504a72d","Type":"ContainerStarted","Data":"7d8f67156c95428502531e2d6d7b0d4ff2b03081bb36d385a3065839f459af50"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.354401 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts"] Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.356589 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g2smk" event={"ID":"5cce043a-2f1b-4f48-967e-c48a00cfe1a6","Type":"ContainerStarted","Data":"8150da0c7bf43ea052c8b9cb2dfefad3eaa9d5a8d1321ad37559f3b798f3d407"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.357153 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.358968 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" event={"ID":"35efd97c-0521-428a-896d-b67490207db5","Type":"ContainerStarted","Data":"103da22d5d7ad80007377bf939401d399bf34ac726213392b0c3e7e78408e88d"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.361872 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" event={"ID":"b189330a-ee63-45f1-8104-4ef173f8ee22","Type":"ContainerStarted","Data":"afd5824e56e3451ab4222828d22b801d69c5d0db9f13ee565fd30c617f1539d4"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.362176 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.363138 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" event={"ID":"e6351c23-e315-4c92-a467-380da403d3c4","Type":"ContainerStarted","Data":"c3711f3be897a53e7f5ed62f4dfea1182824530d785a6c28ea3f9882ad5f65da"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.366582 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" event={"ID":"f8911791-9db1-4463-997e-1ed50da17324","Type":"ContainerStarted","Data":"f3047891deacdaae65ba3e246560c96a6eba06db347aa1943104324a7818b958"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.368351 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" event={"ID":"d8d73604-754b-4dea-9be4-3451964e5589","Type":"ContainerStarted","Data":"ea4ba8929043adf0f17c34319734032d9b4653690dd2e7b618d89cf2bc0f0d30"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.368939 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb"] Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.382245 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpntf\" (UniqueName: \"kubernetes.io/projected/ce7f5025-ebd8-4cbf-af5a-460fbeb681d1-kube-api-access-xpntf\") pod \"dns-default-w7sr4\" (UID: \"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1\") " pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.388504 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" event={"ID":"c9350aae-3053-431e-a2d6-2137f990ca08","Type":"ContainerStarted","Data":"ffa14e1ba61aa831a4df4052a4b8a4a117f7159bef87e57ef471ae2cb78750ae"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.389515 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.389922 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29486880-rhzd2" event={"ID":"00beace7-1e83-40ed-8d92-6da0cae7817f","Type":"ContainerStarted","Data":"79a71111cca31e12c0d06b9a6d103743df145b9a3918dbe29c65faddc57798a0"} Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.449516 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.455018 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:53.955000006 +0000 UTC m=+137.984971007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.494043 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.521548 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.533985 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-spx5f" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.548307 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-w7sr4" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.551619 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.552189 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.052172397 +0000 UTC m=+138.082143398 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.557911 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-64phz" Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.578363 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97"] Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.669160 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.669666 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.169647098 +0000 UTC m=+138.199618099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.770030 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.770147 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.270123305 +0000 UTC m=+138.300094306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.770575 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.770893 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.27088195 +0000 UTC m=+138.300852951 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.792880 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp"] Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.876872 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.877349 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.377331268 +0000 UTC m=+138.407302269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.892526 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz"] Jan 24 00:05:53 crc kubenswrapper[4676]: I0124 00:05:53.980844 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:53 crc kubenswrapper[4676]: E0124 00:05:53.981336 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.481318677 +0000 UTC m=+138.511289678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.041204 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" podStartSLOduration=119.041189654 podStartE2EDuration="1m59.041189654s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:54.008913871 +0000 UTC m=+138.038884872" watchObservedRunningTime="2026-01-24 00:05:54.041189654 +0000 UTC m=+138.071160655" Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.084112 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:54 crc kubenswrapper[4676]: E0124 00:05:54.084741 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.584726038 +0000 UTC m=+138.614697039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.186516 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:54 crc kubenswrapper[4676]: E0124 00:05:54.187105 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.687083995 +0000 UTC m=+138.717054996 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: W0124 00:05:54.203256 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3651bf5e_f692_42b1_8d5e_512daae90cc8.slice/crio-4e0c8c40649393209c1a7f0a24f346edca73c2442af1cad768738ca55f1542e8 WatchSource:0}: Error finding container 4e0c8c40649393209c1a7f0a24f346edca73c2442af1cad768738ca55f1542e8: Status 404 returned error can't find the container with id 4e0c8c40649393209c1a7f0a24f346edca73c2442af1cad768738ca55f1542e8 Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.292544 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:54 crc kubenswrapper[4676]: E0124 00:05:54.292820 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.792799439 +0000 UTC m=+138.822770430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.399712 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:54 crc kubenswrapper[4676]: E0124 00:05:54.400328 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:54.900309102 +0000 UTC m=+138.930280103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.428455 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg"] Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.432113 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ck624"] Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.473738 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29486880-rhzd2" event={"ID":"00beace7-1e83-40ed-8d92-6da0cae7817f","Type":"ContainerStarted","Data":"e742b4dbf395eb8d16eb17141075162ea559f552eabe8302d3372acaca4a7557"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.499998 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" event={"ID":"f8911791-9db1-4463-997e-1ed50da17324","Type":"ContainerStarted","Data":"6d77a0eba3b135c30f1d01ca92806fe520c076b75330c973f7a9495533b64ef2"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.516471 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:54 crc kubenswrapper[4676]: E0124 00:05:54.519903 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.01988567 +0000 UTC m=+139.049856671 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.526815 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-x649d" event={"ID":"32de8698-4bd5-4154-92a3-76930504a72d","Type":"ContainerStarted","Data":"80c19b748f05fd317c9de6c496723038557369b959df5abdea7dec9534b0ced3"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.529673 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.531522 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" event={"ID":"3651bf5e-f692-42b1-8d5e-512daae90cc8","Type":"ContainerStarted","Data":"4e0c8c40649393209c1a7f0a24f346edca73c2442af1cad768738ca55f1542e8"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.555506 4676 generic.go:334] "Generic (PLEG): container finished" podID="35efd97c-0521-428a-896d-b67490207db5" containerID="cb84fc2eb2192156ab29e32f51c5a596bfedf3e62bd0fd8aad07a1f0e3905771" exitCode=0 Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.555845 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" event={"ID":"35efd97c-0521-428a-896d-b67490207db5","Type":"ContainerDied","Data":"cb84fc2eb2192156ab29e32f51c5a596bfedf3e62bd0fd8aad07a1f0e3905771"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.563954 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lgg5l" event={"ID":"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58","Type":"ContainerStarted","Data":"9188b32c88f888426672bc70a99ad5ed649d82e234b0affffed445951bc666a8"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.575907 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kvcv8"] Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.624236 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:54 crc kubenswrapper[4676]: E0124 00:05:54.624856 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.124841611 +0000 UTC m=+139.154812612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.629888 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" event={"ID":"e6351c23-e315-4c92-a467-380da403d3c4","Type":"ContainerStarted","Data":"b6d42d72931ba91bb49ebed4a06d7eef99b1342874b72defa107734f18d238c3"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.637638 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zq745" event={"ID":"7b8f105b-569d-47f2-b564-a0830b010e31","Type":"ContainerStarted","Data":"641343ee68c7d1d57fd397177fd2fd0c1022848fbb6f04704f067ece91ba653b"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.646952 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" event={"ID":"a6360940-ea9d-456d-b546-5a20af404ee5","Type":"ContainerStarted","Data":"779034b47204771a9d445a3c321ee1593f3424dcefc08713ca39e4e19c631ec9"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.649980 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" event={"ID":"aca1e4a5-f702-4803-8f47-7fcb8c7326b6","Type":"ContainerStarted","Data":"f1d0928ddd80044d3a8b61c0eea7a74cedd17dae8f5c908b3c3e078463e661bb"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.650961 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" event={"ID":"977bd2bf-e652-4b16-b8fc-902d4a1d7860","Type":"ContainerStarted","Data":"84f7d28879a217d190a75a520a829f44527ba66ca7e5a7b050285b27c0e379b6"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.662033 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" event={"ID":"9c51c973-c370-41e8-b167-25d3b11418bf","Type":"ContainerStarted","Data":"90417dabd733ff8898e20e70430eced484e20a837f5ba107a83119c85c8c4fc6"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.684338 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" event={"ID":"596a60f1-3eb7-4a57-af4a-1cb1f37f2824","Type":"ContainerStarted","Data":"9467c263be7bed92eb44fc25114a903bc926ef051c31a985c99f3cfd00a7b7ad"} Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.725263 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:54 crc kubenswrapper[4676]: E0124 00:05:54.726395 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.226361261 +0000 UTC m=+139.256332262 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.822242 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29486880-rhzd2" podStartSLOduration=119.8222269 podStartE2EDuration="1m59.8222269s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:54.818856652 +0000 UTC m=+138.848827653" watchObservedRunningTime="2026-01-24 00:05:54.8222269 +0000 UTC m=+138.852197901" Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.829987 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:54 crc kubenswrapper[4676]: E0124 00:05:54.834836 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.334821944 +0000 UTC m=+139.364792945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.866364 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-x649d" podStartSLOduration=119.866345092 podStartE2EDuration="1m59.866345092s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:54.864480073 +0000 UTC m=+138.894451074" watchObservedRunningTime="2026-01-24 00:05:54.866345092 +0000 UTC m=+138.896316093" Jan 24 00:05:54 crc kubenswrapper[4676]: W0124 00:05:54.884624 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8eed212_9137_45e5_8347_1f921fbedb19.slice/crio-7740d756e063c56b647f0d283026bb4d3d2312093aad751bb62487f9b3b5e9ab WatchSource:0}: Error finding container 7740d756e063c56b647f0d283026bb4d3d2312093aad751bb62487f9b3b5e9ab: Status 404 returned error can't find the container with id 7740d756e063c56b647f0d283026bb4d3d2312093aad751bb62487f9b3b5e9ab Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.898863 4676 patch_prober.go:28] interesting pod/console-operator-58897d9998-x649d container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.898911 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-x649d" podUID="32de8698-4bd5-4154-92a3-76930504a72d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/readyz\": dial tcp 10.217.0.25:8443: connect: connection refused" Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.931598 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:54 crc kubenswrapper[4676]: E0124 00:05:54.932124 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.432110799 +0000 UTC m=+139.462081800 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.965589 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" podStartSLOduration=119.96557406 podStartE2EDuration="1m59.96557406s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:54.963275696 +0000 UTC m=+138.993246697" watchObservedRunningTime="2026-01-24 00:05:54.96557406 +0000 UTC m=+138.995545061" Jan 24 00:05:54 crc kubenswrapper[4676]: I0124 00:05:54.966202 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-jmv6d" podStartSLOduration=118.96619833 podStartE2EDuration="1m58.96619833s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:54.922367156 +0000 UTC m=+138.952338157" watchObservedRunningTime="2026-01-24 00:05:54.96619833 +0000 UTC m=+138.996169331" Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.009254 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rkpbj" podStartSLOduration=120.009239638 podStartE2EDuration="2m0.009239638s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:55.008164534 +0000 UTC m=+139.038135535" watchObservedRunningTime="2026-01-24 00:05:55.009239638 +0000 UTC m=+139.039210629" Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.035073 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:55 crc kubenswrapper[4676]: E0124 00:05:55.035458 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.535445107 +0000 UTC m=+139.565416108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.137504 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:55 crc kubenswrapper[4676]: E0124 00:05:55.137900 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.637884976 +0000 UTC m=+139.667855977 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.253136 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:55 crc kubenswrapper[4676]: E0124 00:05:55.253717 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.753706805 +0000 UTC m=+139.783677806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.354217 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:55 crc kubenswrapper[4676]: E0124 00:05:55.354561 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.854538593 +0000 UTC m=+139.884509594 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.466605 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:55 crc kubenswrapper[4676]: E0124 00:05:55.466963 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:55.966946592 +0000 UTC m=+139.996917583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.573867 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:55 crc kubenswrapper[4676]: E0124 00:05:55.574351 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.074333101 +0000 UTC m=+140.104304112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.691863 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:55 crc kubenswrapper[4676]: E0124 00:05:55.692394 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.192352658 +0000 UTC m=+140.222323659 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.734273 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" event={"ID":"d8eed212-9137-45e5-8347-1f921fbedb19","Type":"ContainerStarted","Data":"7740d756e063c56b647f0d283026bb4d3d2312093aad751bb62487f9b3b5e9ab"} Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.778771 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rsw66"] Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.792805 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:55 crc kubenswrapper[4676]: E0124 00:05:55.793422 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.293405754 +0000 UTC m=+140.323376755 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.809831 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" event={"ID":"1eabf562-d289-4685-8ee5-ed1525930d19","Type":"ContainerStarted","Data":"c2eca52027c2d29df7bf8bb00c816d40e324cb64a398fb91764ca8dab5cbaa40"} Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.828206 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" event={"ID":"89435826-645d-48a2-aa3b-f5c42003dcbe","Type":"ContainerStarted","Data":"f8acee532eb12a14f5c8f001c0b4e4969213cd1ade4bba8504365d2ac03f9941"} Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.878582 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" event={"ID":"3651bf5e-f692-42b1-8d5e-512daae90cc8","Type":"ContainerStarted","Data":"162f024bbaf084d516b6f0091d3aba51b6321f7dbfa6359ef5f51639f06926ca"} Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.898833 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:55 crc kubenswrapper[4676]: E0124 00:05:55.899371 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.399357236 +0000 UTC m=+140.429328237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:55 crc kubenswrapper[4676]: W0124 00:05:55.900245 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7af03157_ee92_4e72_a775_acaeabb73e65.slice/crio-3e3f107eb805996a69d5ec1a562a4e16c0b2393227f3f9858f21b04b163f17b3 WatchSource:0}: Error finding container 3e3f107eb805996a69d5ec1a562a4e16c0b2393227f3f9858f21b04b163f17b3: Status 404 returned error can't find the container with id 3e3f107eb805996a69d5ec1a562a4e16c0b2393227f3f9858f21b04b163f17b3 Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.938512 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj"] Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.938555 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb"] Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.938566 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw"] Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.975619 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" event={"ID":"c9350aae-3053-431e-a2d6-2137f990ca08","Type":"ContainerStarted","Data":"ec87b02f17531426f066db608cc81f86442cab14a504474156fb8369b08125a4"} Jan 24 00:05:55 crc kubenswrapper[4676]: I0124 00:05:55.999528 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:55.999876 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.499787522 +0000 UTC m=+140.529758523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.000018 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.001529 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.501512797 +0000 UTC m=+140.531483798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.022503 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-9h9cg" podStartSLOduration=120.022485158 podStartE2EDuration="2m0.022485158s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:56.022391715 +0000 UTC m=+140.052362716" watchObservedRunningTime="2026-01-24 00:05:56.022485158 +0000 UTC m=+140.052456159" Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.048805 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g2smk" event={"ID":"5cce043a-2f1b-4f48-967e-c48a00cfe1a6","Type":"ContainerStarted","Data":"c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250"} Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.106785 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.108018 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.608002757 +0000 UTC m=+140.637973758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.111110 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" event={"ID":"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93","Type":"ContainerStarted","Data":"f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa"} Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.112342 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.135843 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lgg5l" event={"ID":"3e6f77cf-0bf1-4e28-a894-1ca5ee320c58","Type":"ContainerStarted","Data":"c2145c8355a8ea9f0cc532946b4eee599349e33eafdfdf5d910f766bb5ccd739"} Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.162930 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-g2smk" podStartSLOduration=121.162912705 podStartE2EDuration="2m1.162912705s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:56.110943371 +0000 UTC m=+140.140914362" watchObservedRunningTime="2026-01-24 00:05:56.162912705 +0000 UTC m=+140.192883706" Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.163960 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zq745" event={"ID":"7b8f105b-569d-47f2-b564-a0830b010e31","Type":"ContainerStarted","Data":"f0e0c89ffb6bc6404307026b50374c229b53942e5e31b8b6c5141753e3ea8157"} Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.188063 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.209000 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.213701 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.71368652 +0000 UTC m=+140.743657521 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.242897 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" podStartSLOduration=120.242424651 podStartE2EDuration="2m0.242424651s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:56.167713789 +0000 UTC m=+140.197684800" watchObservedRunningTime="2026-01-24 00:05:56.242424651 +0000 UTC m=+140.272395652" Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.243507 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.281198 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-zq745" podStartSLOduration=120.281179581 podStartE2EDuration="2m0.281179581s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:56.274752885 +0000 UTC m=+140.304723886" watchObservedRunningTime="2026-01-24 00:05:56.281179581 +0000 UTC m=+140.311150582" Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.299748 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6m9lm"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.303536 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.315289 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.315635 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.815620974 +0000 UTC m=+140.845591975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.350429 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6kx9j"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.350475 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.355184 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.359441 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-64phz"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.359505 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.359792 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-lgg5l" podStartSLOduration=7.359778238 podStartE2EDuration="7.359778238s" podCreationTimestamp="2026-01-24 00:05:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:56.345747969 +0000 UTC m=+140.375718970" watchObservedRunningTime="2026-01-24 00:05:56.359778238 +0000 UTC m=+140.389749239" Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.433900 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-cdrbc"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.434349 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.434843 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:56.93481125 +0000 UTC m=+140.964782251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: W0124 00:05:56.460207 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4b605a66_7904_4596_a67f_ea21ef41a24b.slice/crio-7ba26b93e116e90166b29605ba52eddeffaddd534a641c9088f052d3d5af4bb7 WatchSource:0}: Error finding container 7ba26b93e116e90166b29605ba52eddeffaddd534a641c9088f052d3d5af4bb7: Status 404 returned error can't find the container with id 7ba26b93e116e90166b29605ba52eddeffaddd534a641c9088f052d3d5af4bb7 Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.491749 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.511620 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.535013 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.535552 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.035532585 +0000 UTC m=+141.065503586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.596611 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:05:56 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:05:56 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:05:56 crc kubenswrapper[4676]: healthz check failed Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.596665 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.640344 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.640773 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.140752134 +0000 UTC m=+141.170723135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.694603 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wvvpg"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.704758 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-spx5f"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.720421 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-w7sr4"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.742633 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.743299 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.243278226 +0000 UTC m=+141.273249227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.744684 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.745034 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.245022802 +0000 UTC m=+141.274993803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.847189 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.848045 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.34801063 +0000 UTC m=+141.377981631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.861455 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.864898 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n8shl"] Jan 24 00:05:56 crc kubenswrapper[4676]: I0124 00:05:56.951812 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:56 crc kubenswrapper[4676]: E0124 00:05:56.952483 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.452469453 +0000 UTC m=+141.482440464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.053151 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.053426 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.553371164 +0000 UTC m=+141.583342165 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.053495 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.054015 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.553998354 +0000 UTC m=+141.583969355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.156813 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.157127 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.657112156 +0000 UTC m=+141.687083157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.191448 4676 patch_prober.go:28] interesting pod/console-operator-58897d9998-x649d container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.191661 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-x649d" podUID="32de8698-4bd5-4154-92a3-76930504a72d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.230596 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" event={"ID":"c2d89ff2-f919-459e-8089-5097aab0f4e2","Type":"ContainerStarted","Data":"0f1131220020d523487c82d70998590e4f63179666319083be7e8c7b5cee3464"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.231428 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-64phz" event={"ID":"290fd523-8c24-458a-8abb-e32ca43caae1","Type":"ContainerStarted","Data":"3fafcb574bc33db2d01bed4dace2b1f846df351963679777f787fad266f7f8bf"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.235599 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" podStartSLOduration=122.235587278 podStartE2EDuration="2m2.235587278s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:57.187424886 +0000 UTC m=+141.217395887" watchObservedRunningTime="2026-01-24 00:05:57.235587278 +0000 UTC m=+141.265558279" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.245876 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" event={"ID":"d8eed212-9137-45e5-8347-1f921fbedb19","Type":"ContainerStarted","Data":"4501b68358fcadcb7cb93da658cb65fc5efa631286b740ec98d8303202c21d47"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.258714 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" event={"ID":"1eabf562-d289-4685-8ee5-ed1525930d19","Type":"ContainerStarted","Data":"572cf18d919a142df3a87fe72606c221700742f64f3b5868d79f863d9965c925"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.267390 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.268040 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.768023257 +0000 UTC m=+141.797994258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.280010 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6m9lm" event={"ID":"4b605a66-7904-4596-a67f-ea21ef41a24b","Type":"ContainerStarted","Data":"7ba26b93e116e90166b29605ba52eddeffaddd534a641c9088f052d3d5af4bb7"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.342088 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" event={"ID":"2f5e70d9-8b16-4684-bd98-4287ccbb6d85","Type":"ContainerStarted","Data":"b2f2d058d7135dc12e594df602d1825fd29a7c58382601f59aebaba322475de2"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.372869 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.373090 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.873036139 +0000 UTC m=+141.903007140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.373343 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.382680 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.882637517 +0000 UTC m=+141.912608518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.477899 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" event={"ID":"8a35921c-db91-45c2-a92c-a42f5b2cab84","Type":"ContainerStarted","Data":"ad68b7c4f8bc12a5320c1bd7f232e02f7a74e0d26a3956a4106d3113d7e10c28"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.478892 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.479295 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:57.979279381 +0000 UTC m=+142.009250382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.506643 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w7sr4" event={"ID":"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1","Type":"ContainerStarted","Data":"3132e83fa488365d9b966d87873edaee96ef6f7c13ad33522cfa1f661dff9e6d"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.515003 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:05:57 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:05:57 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:05:57 crc kubenswrapper[4676]: healthz check failed Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.515044 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.531962 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" podStartSLOduration=122.531943726 podStartE2EDuration="2m2.531943726s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:57.515251762 +0000 UTC m=+141.545222763" watchObservedRunningTime="2026-01-24 00:05:57.531943726 +0000 UTC m=+141.561914717" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.568502 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" event={"ID":"596a60f1-3eb7-4a57-af4a-1cb1f37f2824","Type":"ContainerStarted","Data":"6891fc7226cd73cf2e1d9619ceeaabb04e9b8a541cc964a28a7cdd2d1bc78b99"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.568605 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" event={"ID":"596a60f1-3eb7-4a57-af4a-1cb1f37f2824","Type":"ContainerStarted","Data":"b2f493a2e53381d0f797ca12d2731084d1a59ad8cb2c8880d1dee06227893110"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.568669 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.575564 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" event={"ID":"f53afe6a-307c-4b0d-88cb-596703f35f8a","Type":"ContainerStarted","Data":"29d4f1cd87522474fa72d6fb43bcf59c68668dcfd449d06b68dacbfd7f8808ea"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.579862 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.580363 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:58.080346566 +0000 UTC m=+142.110317567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.592481 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" event={"ID":"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c","Type":"ContainerStarted","Data":"cccadc46c7ef5e3530762fa3243bf0c39b1ea45af73bc9c148e8124f6008387e"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.608110 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" event={"ID":"6d52fce7-049e-441e-8e40-15d044e0319a","Type":"ContainerStarted","Data":"c1bd4c33800e8e92c124a2e705d048651d74ac4490e77bf5314817df6b90d826"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.609771 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" event={"ID":"9d605c02-a15d-46a8-942c-cd85e6ce5452","Type":"ContainerStarted","Data":"bbad16bca0b94c1ecbe3635f858a4da358aad4f756468a9e54016dfc3ee83a53"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.609805 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" event={"ID":"9d605c02-a15d-46a8-942c-cd85e6ce5452","Type":"ContainerStarted","Data":"a59bf3644ff0278dd14e9e4d77ad085e2b7e194b4685e749bf43ab39ae4b8436"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.614482 4676 generic.go:334] "Generic (PLEG): container finished" podID="b189330a-ee63-45f1-8104-4ef173f8ee22" containerID="41953255b89b39f1f003fa2f1dffd6a3957e363ae5d27c9fdb07f020bbead58e" exitCode=0 Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.614542 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" event={"ID":"b189330a-ee63-45f1-8104-4ef173f8ee22","Type":"ContainerDied","Data":"41953255b89b39f1f003fa2f1dffd6a3957e363ae5d27c9fdb07f020bbead58e"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.617857 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" event={"ID":"aca1e4a5-f702-4803-8f47-7fcb8c7326b6","Type":"ContainerStarted","Data":"6172309ae544f7e1556e78a727a310de7dc37c888a8ab8fcbd64f2b2e7c96e2c"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.618658 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.628252 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" event={"ID":"4d53a745-f985-4659-b62c-ce297ce8ce85","Type":"ContainerStarted","Data":"94b32603a4403f69fddee08fd9a7547f2bbb94fb9b6010fd0ea33f0a8816842b"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.629745 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw" event={"ID":"a19dcc7f-c3e9-4aa8-90dc-412550a8060f","Type":"ContainerStarted","Data":"fd121e832fcca06a108857f26f6893e5e3f2b814ea7cb9a7642bb3889987d555"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.631030 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-ml6vn" event={"ID":"d8d73604-754b-4dea-9be4-3451964e5589","Type":"ContainerStarted","Data":"8618f4f528e7015470a790ec69dacfcd1f2c24b04abdd0eb3ba394b12d98b2e8"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.641805 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" podStartSLOduration=121.641794514 podStartE2EDuration="2m1.641794514s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:57.639847092 +0000 UTC m=+141.669818093" watchObservedRunningTime="2026-01-24 00:05:57.641794514 +0000 UTC m=+141.671765515" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.686760 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.687935 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:58.187918591 +0000 UTC m=+142.217889592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.728682 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" event={"ID":"7af03157-ee92-4e72-a775-acaeabb73e65","Type":"ContainerStarted","Data":"a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.728729 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" event={"ID":"7af03157-ee92-4e72-a775-acaeabb73e65","Type":"ContainerStarted","Data":"3e3f107eb805996a69d5ec1a562a4e16c0b2393227f3f9858f21b04b163f17b3"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.729630 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.733199 4676 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rsw66 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.733236 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" podUID="7af03157-ee92-4e72-a775-acaeabb73e65" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.737562 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.749926 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" event={"ID":"25db14a1-725f-42ae-a6e9-646546b584c7","Type":"ContainerStarted","Data":"77972cd045074fba9945778f0832daba3c7a3289379524b5266dd7b755125f15"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.749970 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" event={"ID":"25db14a1-725f-42ae-a6e9-646546b584c7","Type":"ContainerStarted","Data":"d625dcfd007adab1bc03e3fcafd6290f06373efc93d7ae96c3360b773ce74b50"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.788777 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" event={"ID":"35efd97c-0521-428a-896d-b67490207db5","Type":"ContainerStarted","Data":"67c42cb9b12c5cf6127553a26716e03bb1dcb2af2b15485e570076ffaf8c3212"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.788912 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.789187 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:58.289173102 +0000 UTC m=+142.319144103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.790257 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.795216 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" podStartSLOduration=121.795206765 podStartE2EDuration="2m1.795206765s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:57.732113646 +0000 UTC m=+141.762084647" watchObservedRunningTime="2026-01-24 00:05:57.795206765 +0000 UTC m=+141.825177766" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.820202 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" event={"ID":"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e","Type":"ContainerStarted","Data":"faadd71331233346ce0e663f4ccb524437c06a15ad83d6f211042ba07fc44943"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.820245 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" event={"ID":"2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e","Type":"ContainerStarted","Data":"011dff1c3fe125abd38bd940ca07e57dcd72cc95d3be80cb73367df38ffc62bb"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.834207 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" event={"ID":"a6360940-ea9d-456d-b546-5a20af404ee5","Type":"ContainerStarted","Data":"05f13459a2f3b055c50ed252a99d98ca75b4054ee677300bd00fe6e1ff08d44a"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.856803 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2xb97" podStartSLOduration=121.856784177 podStartE2EDuration="2m1.856784177s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:57.797130098 +0000 UTC m=+141.827101099" watchObservedRunningTime="2026-01-24 00:05:57.856784177 +0000 UTC m=+141.886755178" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.868674 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" podStartSLOduration=122.868651087 podStartE2EDuration="2m2.868651087s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:57.851819418 +0000 UTC m=+141.881790429" watchObservedRunningTime="2026-01-24 00:05:57.868651087 +0000 UTC m=+141.898622088" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.882228 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" podStartSLOduration=121.882209361 podStartE2EDuration="2m1.882209361s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:57.881141197 +0000 UTC m=+141.911112198" watchObservedRunningTime="2026-01-24 00:05:57.882209361 +0000 UTC m=+141.912180362" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.886619 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" event={"ID":"3651bf5e-f692-42b1-8d5e-512daae90cc8","Type":"ContainerStarted","Data":"77ede6dcf71f40244a89448f3c333530ae3535f093766fe2f7cada2a7b8f1805"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.892711 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.894037 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:58.394019679 +0000 UTC m=+142.423990680 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.908961 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" event={"ID":"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106","Type":"ContainerStarted","Data":"7b39dfb3fd225ef1979265fdd2e45cfe4514080b41db9b075f6fa61cedd4f6e2"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.968334 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" event={"ID":"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1","Type":"ContainerStarted","Data":"48552c7ac7205c236a824947a2c6d70c82f2be4c38380e4a2ff3a395cef9bd8b"} Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.969524 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.970475 4676 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-r6r59 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.970521 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" podUID="fdb046fc-eba9-4f07-a1d1-2db71a2d46c1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Jan 24 00:05:57 crc kubenswrapper[4676]: I0124 00:05:57.993706 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:57 crc kubenswrapper[4676]: E0124 00:05:57.994943 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:58.49491469 +0000 UTC m=+142.524885691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.016313 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" event={"ID":"ae793283-426a-4e02-b96b-89c3f16f2d16","Type":"ContainerStarted","Data":"278e06bc52d2c5dd807374de718b0b08159c45c4ffe1444c4ceca51bb82969ec"} Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.017772 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" event={"ID":"c50241f8-d135-4df8-b047-e76fb28b8a3d","Type":"ContainerStarted","Data":"9b58dea0bbca8fdb9cff011940784b3be0d4f2e9f630c86d3a6b54ff117e955b"} Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.018547 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.019706 4676 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-2cgk2 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.019751 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" podUID="c50241f8-d135-4df8-b047-e76fb28b8a3d" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.030771 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" event={"ID":"89435826-645d-48a2-aa3b-f5c42003dcbe","Type":"ContainerStarted","Data":"b436a79b5d8ef5e9fda6092960f14ec8555b8c360d001d483efb56a39a863e65"} Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.031749 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.058360 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" event={"ID":"977bd2bf-e652-4b16-b8fc-902d4a1d7860","Type":"ContainerStarted","Data":"336fa4865cfc985bea40eacba6d369a0e38190209e4cc96e7fbeda326c923f35"} Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.102739 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zc8ts" podStartSLOduration=123.102720771 podStartE2EDuration="2m3.102720771s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:58.01147255 +0000 UTC m=+142.041443541" watchObservedRunningTime="2026-01-24 00:05:58.102720771 +0000 UTC m=+142.132691772" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.105072 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:58 crc kubenswrapper[4676]: E0124 00:05:58.106200 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:58.606186802 +0000 UTC m=+142.636157803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.127147 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-spx5f" event={"ID":"f5c752b5-1413-4c4e-aea6-5302a7c69467","Type":"ContainerStarted","Data":"8b99ff8354c09cbd893a20e89d11dda139e4d7edf9f2b97679080aaa79aa03cc"} Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.208291 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:58 crc kubenswrapper[4676]: E0124 00:05:58.216990 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:58.71697565 +0000 UTC m=+142.746946651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.263155 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6wjpz" podStartSLOduration=122.263130087 podStartE2EDuration="2m2.263130087s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:58.10455325 +0000 UTC m=+142.134524251" watchObservedRunningTime="2026-01-24 00:05:58.263130087 +0000 UTC m=+142.293101088" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.312626 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:58 crc kubenswrapper[4676]: E0124 00:05:58.313224 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:58.81320427 +0000 UTC m=+142.843175271 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.410585 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" podStartSLOduration=123.410569607 podStartE2EDuration="2m3.410569607s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:58.26604921 +0000 UTC m=+142.296020211" watchObservedRunningTime="2026-01-24 00:05:58.410569607 +0000 UTC m=+142.440540608" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.411209 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" podStartSLOduration=122.411205008 podStartE2EDuration="2m2.411205008s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:58.409894676 +0000 UTC m=+142.439865677" watchObservedRunningTime="2026-01-24 00:05:58.411205008 +0000 UTC m=+142.441176009" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.414061 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:58 crc kubenswrapper[4676]: E0124 00:05:58.414428 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:58.914415731 +0000 UTC m=+142.944386732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.514630 4676 csr.go:261] certificate signing request csr-gztrj is approved, waiting to be issued Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.515629 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:58 crc kubenswrapper[4676]: E0124 00:05:58.516012 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.015997493 +0000 UTC m=+143.045968484 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.524697 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:05:58 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:05:58 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:05:58 crc kubenswrapper[4676]: healthz check failed Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.524760 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.547838 4676 csr.go:257] certificate signing request csr-gztrj is issued Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.556006 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" podStartSLOduration=122.555980864 podStartE2EDuration="2m2.555980864s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:58.471256131 +0000 UTC m=+142.501227132" watchObservedRunningTime="2026-01-24 00:05:58.555980864 +0000 UTC m=+142.585951865" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.617648 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:58 crc kubenswrapper[4676]: E0124 00:05:58.617983 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.117971848 +0000 UTC m=+143.147942849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.652006 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-4nnjp" podStartSLOduration=122.651992827 podStartE2EDuration="2m2.651992827s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:58.572977917 +0000 UTC m=+142.602948918" watchObservedRunningTime="2026-01-24 00:05:58.651992827 +0000 UTC m=+142.681963828" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.725489 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:58 crc kubenswrapper[4676]: E0124 00:05:58.725905 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.225889153 +0000 UTC m=+143.255860154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.827028 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:58 crc kubenswrapper[4676]: E0124 00:05:58.827792 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.327780195 +0000 UTC m=+143.357751196 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.852624 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" podStartSLOduration=123.85260418 podStartE2EDuration="2m3.85260418s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:58.656657967 +0000 UTC m=+142.686628968" watchObservedRunningTime="2026-01-24 00:05:58.85260418 +0000 UTC m=+142.882575181" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.853191 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lxs47"] Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.854035 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.860636 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.872531 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lxs47"] Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.928981 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.929177 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-utilities\") pod \"certified-operators-lxs47\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.929205 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjhrt\" (UniqueName: \"kubernetes.io/projected/592859d8-1f7e-4e35-acc9-635e130ad2d2-kube-api-access-tjhrt\") pod \"certified-operators-lxs47\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:58 crc kubenswrapper[4676]: I0124 00:05:58.929229 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-catalog-content\") pod \"certified-operators-lxs47\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:58 crc kubenswrapper[4676]: E0124 00:05:58.929394 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.429361437 +0000 UTC m=+143.459332438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.030621 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjhrt\" (UniqueName: \"kubernetes.io/projected/592859d8-1f7e-4e35-acc9-635e130ad2d2-kube-api-access-tjhrt\") pod \"certified-operators-lxs47\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.030904 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-catalog-content\") pod \"certified-operators-lxs47\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.030956 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.031023 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-utilities\") pod \"certified-operators-lxs47\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.031365 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.531348703 +0000 UTC m=+143.561319704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.031396 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-utilities\") pod \"certified-operators-lxs47\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.031608 4676 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-kvcv8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.031653 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" podUID="89435826-645d-48a2-aa3b-f5c42003dcbe" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.031770 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-catalog-content\") pod \"certified-operators-lxs47\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.062598 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rxmrn"] Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.063444 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.070324 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.084181 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rxmrn"] Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.109152 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjhrt\" (UniqueName: \"kubernetes.io/projected/592859d8-1f7e-4e35-acc9-635e130ad2d2-kube-api-access-tjhrt\") pod \"certified-operators-lxs47\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.131615 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.131809 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-utilities\") pod \"community-operators-rxmrn\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.131840 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzqlz\" (UniqueName: \"kubernetes.io/projected/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-kube-api-access-hzqlz\") pod \"community-operators-rxmrn\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.131873 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-catalog-content\") pod \"community-operators-rxmrn\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.131997 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.631983975 +0000 UTC m=+143.661954976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.168742 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" event={"ID":"ae793283-426a-4e02-b96b-89c3f16f2d16","Type":"ContainerStarted","Data":"ee9da2b3e0f65e35ab572af64682e924c0f611a793e068bb8c18a0ee782bc896"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.173625 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.185235 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-wvvpg" podStartSLOduration=124.185222159 podStartE2EDuration="2m4.185222159s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.184125984 +0000 UTC m=+143.214096985" watchObservedRunningTime="2026-01-24 00:05:59.185222159 +0000 UTC m=+143.215193150" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.208077 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" event={"ID":"25db14a1-725f-42ae-a6e9-646546b584c7","Type":"ContainerStarted","Data":"f248b0348835be7d1147fcd695dd3066767299eda9ea4ffd9ff373a3f2b4259c"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.238650 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-utilities\") pod \"community-operators-rxmrn\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.238695 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzqlz\" (UniqueName: \"kubernetes.io/projected/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-kube-api-access-hzqlz\") pod \"community-operators-rxmrn\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.238731 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.238766 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-catalog-content\") pod \"community-operators-rxmrn\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.240692 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.740678485 +0000 UTC m=+143.770649486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.241199 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-utilities\") pod \"community-operators-rxmrn\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.244445 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-catalog-content\") pod \"community-operators-rxmrn\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.245060 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" event={"ID":"8a35921c-db91-45c2-a92c-a42f5b2cab84","Type":"ContainerStarted","Data":"d18eed4d040a1801f33a1a546fb0c31a050f213bcf9c3be8243646f6c7b8f091"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.246386 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-spx5f" event={"ID":"f5c752b5-1413-4c4e-aea6-5302a7c69467","Type":"ContainerStarted","Data":"e943b53f8d0728d98dcdd6d0078374e761cb1c0dae7bf0d94025557ea8b3bad7"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.262407 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" event={"ID":"e3f3c4fe-f6a3-4772-9551-3a0cf1bd5106","Type":"ContainerStarted","Data":"bf973003ab60349a19acd5d035e6d0d4442245f260794c5e37916a0c4c853df5"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.272701 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6kx9j" event={"ID":"4d53a745-f985-4659-b62c-ce297ce8ce85","Type":"ContainerStarted","Data":"ef308860d3684c8dcf0bcbf34ed1bc677d571ced579648c5ee47df26b4fe9bb3"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.285880 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wh9bw" podStartSLOduration=123.285859971 podStartE2EDuration="2m3.285859971s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.244298121 +0000 UTC m=+143.274269122" watchObservedRunningTime="2026-01-24 00:05:59.285859971 +0000 UTC m=+143.315830972" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.287562 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9lq6l"] Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.300139 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzqlz\" (UniqueName: \"kubernetes.io/projected/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-kube-api-access-hzqlz\") pod \"community-operators-rxmrn\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.300511 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.301113 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lq6l"] Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.312453 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw" event={"ID":"a19dcc7f-c3e9-4aa8-90dc-412550a8060f","Type":"ContainerStarted","Data":"66ee604da142844888fae6dbb5d879f5a590502f3e8e6b2d15831781c77e7ece"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.312493 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw" event={"ID":"a19dcc7f-c3e9-4aa8-90dc-412550a8060f","Type":"ContainerStarted","Data":"a90652cb8dbf7a8ff274bf2602ccbe67a8818f9961b942e19bab908916510b4f"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.322862 4676 generic.go:334] "Generic (PLEG): container finished" podID="1eabf562-d289-4685-8ee5-ed1525930d19" containerID="572cf18d919a142df3a87fe72606c221700742f64f3b5868d79f863d9965c925" exitCode=0 Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.323037 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" event={"ID":"1eabf562-d289-4685-8ee5-ed1525930d19","Type":"ContainerDied","Data":"572cf18d919a142df3a87fe72606c221700742f64f3b5868d79f863d9965c925"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.323646 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-spx5f" podStartSLOduration=10.32362349 podStartE2EDuration="10.32362349s" podCreationTimestamp="2026-01-24 00:05:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.298099613 +0000 UTC m=+143.328070614" watchObservedRunningTime="2026-01-24 00:05:59.32362349 +0000 UTC m=+143.353594481" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.341441 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.341635 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdl5n\" (UniqueName: \"kubernetes.io/projected/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-kube-api-access-jdl5n\") pod \"certified-operators-9lq6l\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.341663 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-catalog-content\") pod \"certified-operators-9lq6l\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.341746 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-utilities\") pod \"certified-operators-9lq6l\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.345007 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.844988685 +0000 UTC m=+143.874959686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.385752 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" event={"ID":"fdb046fc-eba9-4f07-a1d1-2db71a2d46c1","Type":"ContainerStarted","Data":"6693a80f134e8c29e380a37d9d451ccb5bd9dbeb65aabe0df8490152f17f06ce"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.426120 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-q8ljj" podStartSLOduration=124.426101822 podStartE2EDuration="2m4.426101822s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.33268871 +0000 UTC m=+143.362659711" watchObservedRunningTime="2026-01-24 00:05:59.426101822 +0000 UTC m=+143.456072823" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.432010 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.445670 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdl5n\" (UniqueName: \"kubernetes.io/projected/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-kube-api-access-jdl5n\") pod \"certified-operators-9lq6l\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.445731 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-catalog-content\") pod \"certified-operators-9lq6l\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.445834 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.445928 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-utilities\") pod \"certified-operators-9lq6l\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.454596 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-catalog-content\") pod \"certified-operators-9lq6l\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.456146 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:05:59.956121943 +0000 UTC m=+143.986092944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.458089 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-utilities\") pod \"certified-operators-9lq6l\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.489967 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdl5n\" (UniqueName: \"kubernetes.io/projected/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-kube-api-access-jdl5n\") pod \"certified-operators-9lq6l\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.490081 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pnqjd"] Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.545013 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.546571 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.546697 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-catalog-content\") pod \"community-operators-pnqjd\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.546759 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-utilities\") pod \"community-operators-pnqjd\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.546795 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49sw6\" (UniqueName: \"kubernetes.io/projected/891b78f7-509c-4e8d-b846-52881396a64d-kube-api-access-49sw6\") pod \"community-operators-pnqjd\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.546890 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.046877159 +0000 UTC m=+144.076848160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.548959 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-24 00:00:58 +0000 UTC, rotation deadline is 2026-11-25 01:57:09.832515001 +0000 UTC Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.549012 4676 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7321h51m10.283506584s for next certificate rotation Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.555516 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hjbqw" podStartSLOduration=123.547176167 podStartE2EDuration="2m3.547176167s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.471291369 +0000 UTC m=+143.501262390" watchObservedRunningTime="2026-01-24 00:05:59.547176167 +0000 UTC m=+143.577147168" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.561801 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pnqjd"] Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.561852 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" event={"ID":"9d605c02-a15d-46a8-942c-cd85e6ce5452","Type":"ContainerStarted","Data":"b02350aa3570b328b0bcc70e7d67adfee447748aa0c39b31cbf01d8a91ea118e"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.566793 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:05:59 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:05:59 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:05:59 crc kubenswrapper[4676]: healthz check failed Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.566846 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.585073 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" event={"ID":"b189330a-ee63-45f1-8104-4ef173f8ee22","Type":"ContainerStarted","Data":"96d356ad0380f380a69eff524b4f1dc9c2ba310f7ae2f53d7cb02846df923de8"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.601041 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6m9lm" event={"ID":"4b605a66-7904-4596-a67f-ea21ef41a24b","Type":"ContainerStarted","Data":"3edd92bd702125f593d90a1fbce6f0f738ec4a56e64db7b21757f92df15eb7cf"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.601927 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6m9lm" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.603038 4676 patch_prober.go:28] interesting pod/downloads-7954f5f757-6m9lm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.603075 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6m9lm" podUID="4b605a66-7904-4596-a67f-ea21ef41a24b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.604604 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w7sr4" event={"ID":"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1","Type":"ContainerStarted","Data":"8bf3837f26c124fb3dea1cb5af5423060dde2aad405ee85078a315680369e231"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.609351 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" event={"ID":"c2d89ff2-f919-459e-8089-5097aab0f4e2","Type":"ContainerStarted","Data":"58b1c728726962279da6279af62f496118daac9c9e348c61cab85e7d8d43f7c5"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.622966 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" event={"ID":"c50241f8-d135-4df8-b047-e76fb28b8a3d","Type":"ContainerStarted","Data":"e7ce28b323b8b8486c8ab8d191dde3eb75dc5db6d7b9860b5ba5c6573304aea7"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.625037 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-nhmlj" podStartSLOduration=123.625022551 podStartE2EDuration="2m3.625022551s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.620069341 +0000 UTC m=+143.650040342" watchObservedRunningTime="2026-01-24 00:05:59.625022551 +0000 UTC m=+143.654993552" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.648034 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.648190 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-catalog-content\") pod \"community-operators-pnqjd\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.648290 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-utilities\") pod \"community-operators-pnqjd\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.648356 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49sw6\" (UniqueName: \"kubernetes.io/projected/891b78f7-509c-4e8d-b846-52881396a64d-kube-api-access-49sw6\") pod \"community-operators-pnqjd\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.649790 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.149772143 +0000 UTC m=+144.179743144 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.653875 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-catalog-content\") pod \"community-operators-pnqjd\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.655784 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-utilities\") pod \"community-operators-pnqjd\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.659023 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" event={"ID":"6d52fce7-049e-441e-8e40-15d044e0319a","Type":"ContainerStarted","Data":"1d08010e79a9253ef911ab9df7ef942a3cfa181b9ed3fadcb6ae76c24d32cbaf"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.659060 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" event={"ID":"6d52fce7-049e-441e-8e40-15d044e0319a","Type":"ContainerStarted","Data":"51af42e11a2cb3b8c0b0d79b6aa83a9b73beada8a6246f775c3f887692bd6a5a"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.673927 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" event={"ID":"d8eed212-9137-45e5-8347-1f921fbedb19","Type":"ContainerStarted","Data":"b35684c3824b37433ff463a7c5ce25ec374136425ad48494c561190706b6df5d"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.688903 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" event={"ID":"2f5e70d9-8b16-4684-bd98-4287ccbb6d85","Type":"ContainerStarted","Data":"8460fe10f124ae6d86e08e1e7bb4e37f2b62c4cc301709f8e03e0a378432f4b2"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.705147 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-2cgk2" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.719144 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.732244 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" podStartSLOduration=123.732205022 podStartE2EDuration="2m3.732205022s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.679113272 +0000 UTC m=+143.709084273" watchObservedRunningTime="2026-01-24 00:05:59.732205022 +0000 UTC m=+143.762176023" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.734769 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49sw6\" (UniqueName: \"kubernetes.io/projected/891b78f7-509c-4e8d-b846-52881396a64d-kube-api-access-49sw6\") pod \"community-operators-pnqjd\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.735499 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xt6kb" podStartSLOduration=124.735491087 podStartE2EDuration="2m4.735491087s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.731397696 +0000 UTC m=+143.761368697" watchObservedRunningTime="2026-01-24 00:05:59.735491087 +0000 UTC m=+143.765462088" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.743209 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" event={"ID":"f53afe6a-307c-4b0d-88cb-596703f35f8a","Type":"ContainerStarted","Data":"a2a46bead832ce59b31551b91c40dbbcccf807b31102af079db87e5f92879d21"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.754978 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.755160 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.255142146 +0000 UTC m=+144.285113147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.755759 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.756067 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.256055486 +0000 UTC m=+144.286026487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.775271 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ck624" podStartSLOduration=123.77525322 podStartE2EDuration="2m3.77525322s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.774461485 +0000 UTC m=+143.804432486" watchObservedRunningTime="2026-01-24 00:05:59.77525322 +0000 UTC m=+143.805224211" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.789783 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" event={"ID":"bf0662c1-a2ab-4bc6-8278-5c6cbb43cf2c","Type":"ContainerStarted","Data":"15c98b0b70967786e363fc8732cd1aebce0f294589ac539df0ca5c027ee5ec22"} Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.791141 4676 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rsw66 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.791200 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" podUID="7af03157-ee92-4e72-a775-acaeabb73e65" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.828894 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-sxjwn" podStartSLOduration=123.819556598 podStartE2EDuration="2m3.819556598s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.818097692 +0000 UTC m=+143.848068693" watchObservedRunningTime="2026-01-24 00:05:59.819556598 +0000 UTC m=+143.849527600" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.860994 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6m9lm" podStartSLOduration=124.860967884 podStartE2EDuration="2m4.860967884s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.858823136 +0000 UTC m=+143.888794137" watchObservedRunningTime="2026-01-24 00:05:59.860967884 +0000 UTC m=+143.890938875" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.868987 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.870733 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.370709817 +0000 UTC m=+144.400680818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.886829 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.897320 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.397280817 +0000 UTC m=+144.427251818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.937522 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-mkwj8" podStartSLOduration=123.937477704 podStartE2EDuration="2m3.937477704s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.885656525 +0000 UTC m=+143.915627526" watchObservedRunningTime="2026-01-24 00:05:59.937477704 +0000 UTC m=+143.967448705" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.946755 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:05:59 crc kubenswrapper[4676]: I0124 00:05:59.990141 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:05:59 crc kubenswrapper[4676]: E0124 00:05:59.990732 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.490707908 +0000 UTC m=+144.520678909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.060272 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n8shl" podStartSLOduration=124.060244354 podStartE2EDuration="2m4.060244354s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:06:00.054127168 +0000 UTC m=+144.084098169" watchObservedRunningTime="2026-01-24 00:06:00.060244354 +0000 UTC m=+144.090215355" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.061067 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-975pl" podStartSLOduration=124.061059841 podStartE2EDuration="2m4.061059841s" podCreationTimestamp="2026-01-24 00:03:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:05:59.977818486 +0000 UTC m=+144.007789487" watchObservedRunningTime="2026-01-24 00:06:00.061059841 +0000 UTC m=+144.091030842" Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.092552 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.592301321 +0000 UTC m=+144.622272322 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.091837 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.142058 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-7zllh" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.194260 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.194810 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.694793352 +0000 UTC m=+144.724764353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.195115 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.195643 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.69563569 +0000 UTC m=+144.725606691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.297978 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.298397 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.798367429 +0000 UTC m=+144.828338430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.392450 4676 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-r6r59 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.392532 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" podUID="fdb046fc-eba9-4f07-a1d1-2db71a2d46c1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.402766 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.403345 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:00.90332847 +0000 UTC m=+144.933299471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.508330 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.509059 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.009043194 +0000 UTC m=+145.039014195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.515672 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:00 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:00 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:00 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.515723 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.612092 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.612487 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.112475756 +0000 UTC m=+145.142446747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.714860 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.715332 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.215315038 +0000 UTC m=+145.245286029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.794503 4676 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-kvcv8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.794583 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" podUID="89435826-645d-48a2-aa3b-f5c42003dcbe" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.814741 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w7sr4" event={"ID":"ce7f5025-ebd8-4cbf-af5a-460fbeb681d1","Type":"ContainerStarted","Data":"6711c3613b3b509231fd33b0dc5c3e7bb7bcceb4c5432073d4be1891e4207a35"} Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.814829 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-w7sr4" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.819085 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.819425 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.319413751 +0000 UTC m=+145.349384752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.822837 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-64phz" event={"ID":"290fd523-8c24-458a-8abb-e32ca43caae1","Type":"ContainerStarted","Data":"6bf62bc4a416ff0bd0b8789a0d98c2c4fda57ab1304842586447edb37f969335"} Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.834228 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" event={"ID":"8a35921c-db91-45c2-a92c-a42f5b2cab84","Type":"ContainerStarted","Data":"96bf3915a948d562927938299e55bac7c30f343e96c87a15a7fa3a3f486b1edb"} Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.834617 4676 patch_prober.go:28] interesting pod/downloads-7954f5f757-6m9lm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.834649 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6m9lm" podUID="4b605a66-7904-4596-a67f-ea21ef41a24b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.839253 4676 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rsw66 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" start-of-body= Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.839282 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" podUID="7af03157-ee92-4e72-a775-acaeabb73e65" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.34:8080/healthz\": dial tcp 10.217.0.34:8080: connect: connection refused" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.924600 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:00 crc kubenswrapper[4676]: E0124 00:06:00.928040 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.428022128 +0000 UTC m=+145.457993129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.984171 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-w7sr4" podStartSLOduration=10.984152065 podStartE2EDuration="10.984152065s" podCreationTimestamp="2026-01-24 00:05:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:06:00.927605235 +0000 UTC m=+144.957576236" watchObservedRunningTime="2026-01-24 00:06:00.984152065 +0000 UTC m=+145.014123066" Jan 24 00:06:00 crc kubenswrapper[4676]: I0124 00:06:00.985128 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-cdrbc" podStartSLOduration=125.985122817 podStartE2EDuration="2m5.985122817s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:06:00.983757882 +0000 UTC m=+145.013728883" watchObservedRunningTime="2026-01-24 00:06:00.985122817 +0000 UTC m=+145.015093818" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.029432 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.037935 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.537913447 +0000 UTC m=+145.567884448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.051149 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lxs47"] Jan 24 00:06:01 crc kubenswrapper[4676]: W0124 00:06:01.069077 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod592859d8_1f7e_4e35_acc9_635e130ad2d2.slice/crio-13ab7acc7e8b9f7cc9283df9a7fe0c7c12a31d5e17ec678008c5409224450c7c WatchSource:0}: Error finding container 13ab7acc7e8b9f7cc9283df9a7fe0c7c12a31d5e17ec678008c5409224450c7c: Status 404 returned error can't find the container with id 13ab7acc7e8b9f7cc9283df9a7fe0c7c12a31d5e17ec678008c5409224450c7c Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.100876 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l7nsm"] Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.101967 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.109969 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.130787 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.131205 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.631186663 +0000 UTC m=+145.661157664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.163422 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7nsm"] Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.234061 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.234103 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-utilities\") pod \"redhat-marketplace-l7nsm\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.234150 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-catalog-content\") pod \"redhat-marketplace-l7nsm\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.234172 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4p5l\" (UniqueName: \"kubernetes.io/projected/920c325a-f36b-4162-9d37-ea88124be938-kube-api-access-n4p5l\") pod \"redhat-marketplace-l7nsm\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.234462 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.73445153 +0000 UTC m=+145.764422531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.318774 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rxmrn"] Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.335874 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.336010 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-catalog-content\") pod \"redhat-marketplace-l7nsm\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.336031 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4p5l\" (UniqueName: \"kubernetes.io/projected/920c325a-f36b-4162-9d37-ea88124be938-kube-api-access-n4p5l\") pod \"redhat-marketplace-l7nsm\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.336116 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-utilities\") pod \"redhat-marketplace-l7nsm\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.336871 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-utilities\") pod \"redhat-marketplace-l7nsm\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.336932 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.8369187 +0000 UTC m=+145.866889701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.337121 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-catalog-content\") pod \"redhat-marketplace-l7nsm\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.448143 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.448718 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:01.948703729 +0000 UTC m=+145.978674730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.449616 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4p5l\" (UniqueName: \"kubernetes.io/projected/920c325a-f36b-4162-9d37-ea88124be938-kube-api-access-n4p5l\") pod \"redhat-marketplace-l7nsm\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.454676 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-94fls"] Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.462890 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.486661 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-94fls"] Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.505003 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.512763 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:01 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:01 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:01 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.512805 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.550854 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.551116 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5djr\" (UniqueName: \"kubernetes.io/projected/236fa0ff-93a4-429d-9a2a-b1ae84167818-kube-api-access-g5djr\") pod \"redhat-marketplace-94fls\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.551251 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-utilities\") pod \"redhat-marketplace-94fls\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.551370 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.051271813 +0000 UTC m=+146.081242804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.551587 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-catalog-content\") pod \"redhat-marketplace-94fls\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.551715 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.552254 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.052245774 +0000 UTC m=+146.082216775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.654805 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.655221 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5djr\" (UniqueName: \"kubernetes.io/projected/236fa0ff-93a4-429d-9a2a-b1ae84167818-kube-api-access-g5djr\") pod \"redhat-marketplace-94fls\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.655409 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-utilities\") pod \"redhat-marketplace-94fls\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.655564 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-catalog-content\") pod \"redhat-marketplace-94fls\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.656085 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-catalog-content\") pod \"redhat-marketplace-94fls\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.656250 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.156231853 +0000 UTC m=+146.186202854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.657057 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-utilities\") pod \"redhat-marketplace-94fls\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.704742 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lq6l"] Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.707232 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5djr\" (UniqueName: \"kubernetes.io/projected/236fa0ff-93a4-429d-9a2a-b1ae84167818-kube-api-access-g5djr\") pod \"redhat-marketplace-94fls\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.756956 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.757289 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.257277368 +0000 UTC m=+146.287248369 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.779422 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pnqjd"] Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.786713 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.834669 4676 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-kvcv8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.834809 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" podUID="89435826-645d-48a2-aa3b-f5c42003dcbe" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.838216 4676 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-r6r59 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.838267 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" podUID="fdb046fc-eba9-4f07-a1d1-2db71a2d46c1" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.858848 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.859027 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.359011476 +0000 UTC m=+146.388982477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.859210 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.859550 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.359530822 +0000 UTC m=+146.389501823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.861058 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lq6l" event={"ID":"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59","Type":"ContainerStarted","Data":"240a15f7e6730a7c6ad0337131f8a5c437c590dd34ae29a62a731e625e9f6839"} Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.864169 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxmrn" event={"ID":"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd","Type":"ContainerStarted","Data":"da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060"} Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.864201 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxmrn" event={"ID":"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd","Type":"ContainerStarted","Data":"e7b1f0797950f02a9f5633d0b1bb88fc987a1236bb934934be7f6364e391819f"} Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.868797 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.868827 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.900321 4676 patch_prober.go:28] interesting pod/apiserver-76f77b778f-2scbc container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]log ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]etcd ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/generic-apiserver-start-informers ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/max-in-flight-filter ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 24 00:06:01 crc kubenswrapper[4676]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/project.openshift.io-projectcache ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/openshift.io-startinformers ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 24 00:06:01 crc kubenswrapper[4676]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 24 00:06:01 crc kubenswrapper[4676]: livez check failed Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.900407 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" podUID="2f4cf2ff-5d1d-4d40-b7ca-02d8d681e70e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.901716 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" event={"ID":"1eabf562-d289-4685-8ee5-ed1525930d19","Type":"ContainerDied","Data":"c2eca52027c2d29df7bf8bb00c816d40e324cb64a398fb91764ca8dab5cbaa40"} Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.901755 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2eca52027c2d29df7bf8bb00c816d40e324cb64a398fb91764ca8dab5cbaa40" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.901836 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.902171 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.939312 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxs47" event={"ID":"592859d8-1f7e-4e35-acc9-635e130ad2d2","Type":"ContainerStarted","Data":"5f55960a6c743513af98906e24b7fac1e853e9cbe06b95180d6ac6353a01cca4"} Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.939359 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxs47" event={"ID":"592859d8-1f7e-4e35-acc9-635e130ad2d2","Type":"ContainerStarted","Data":"13ab7acc7e8b9f7cc9283df9a7fe0c7c12a31d5e17ec678008c5409224450c7c"} Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.945255 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-64phz" event={"ID":"290fd523-8c24-458a-8abb-e32ca43caae1","Type":"ContainerStarted","Data":"2f417b8a42114b9e1c7520500e311fcb7c0e793e5961df975148a4209f2eb997"} Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.945855 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.960241 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eabf562-d289-4685-8ee5-ed1525930d19-secret-volume\") pod \"1eabf562-d289-4685-8ee5-ed1525930d19\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.960468 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.960525 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eabf562-d289-4685-8ee5-ed1525930d19-config-volume\") pod \"1eabf562-d289-4685-8ee5-ed1525930d19\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.960582 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvj8q\" (UniqueName: \"kubernetes.io/projected/1eabf562-d289-4685-8ee5-ed1525930d19-kube-api-access-gvj8q\") pod \"1eabf562-d289-4685-8ee5-ed1525930d19\" (UID: \"1eabf562-d289-4685-8ee5-ed1525930d19\") " Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.964139 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnqjd" event={"ID":"891b78f7-509c-4e8d-b846-52881396a64d","Type":"ContainerStarted","Data":"fe1008d35989eb8d8f31d34ea59eda3fc8a16f8472a8718f92a963e88a25a092"} Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.965529 4676 patch_prober.go:28] interesting pod/downloads-7954f5f757-6m9lm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.965578 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6m9lm" podUID="4b605a66-7904-4596-a67f-ea21ef41a24b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Jan 24 00:06:01 crc kubenswrapper[4676]: E0124 00:06:01.966038 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.466016681 +0000 UTC m=+146.495987692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.969953 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eabf562-d289-4685-8ee5-ed1525930d19-config-volume" (OuterVolumeSpecName: "config-volume") pod "1eabf562-d289-4685-8ee5-ed1525930d19" (UID: "1eabf562-d289-4685-8ee5-ed1525930d19"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.970318 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-x649d" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.974723 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eabf562-d289-4685-8ee5-ed1525930d19-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1eabf562-d289-4685-8ee5-ed1525930d19" (UID: "1eabf562-d289-4685-8ee5-ed1525930d19"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:06:01 crc kubenswrapper[4676]: I0124 00:06:01.992837 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eabf562-d289-4685-8ee5-ed1525930d19-kube-api-access-gvj8q" (OuterVolumeSpecName: "kube-api-access-gvj8q") pod "1eabf562-d289-4685-8ee5-ed1525930d19" (UID: "1eabf562-d289-4685-8ee5-ed1525930d19"). InnerVolumeSpecName "kube-api-access-gvj8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.031139 4676 patch_prober.go:28] interesting pod/console-f9d7485db-g2smk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.031370 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-g2smk" podUID="5cce043a-2f1b-4f48-967e-c48a00cfe1a6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.031133 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.031681 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.065439 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.065876 4676 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eabf562-d289-4685-8ee5-ed1525930d19-config-volume\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.066007 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvj8q\" (UniqueName: \"kubernetes.io/projected/1eabf562-d289-4685-8ee5-ed1525930d19-kube-api-access-gvj8q\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.066116 4676 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eabf562-d289-4685-8ee5-ed1525930d19-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.067285 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.567264883 +0000 UTC m=+146.597235884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.094524 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.094743 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.115795 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-645jr"] Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.116912 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eabf562-d289-4685-8ee5-ed1525930d19" containerName="collect-profiles" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.117102 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eabf562-d289-4685-8ee5-ed1525930d19" containerName="collect-profiles" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.117283 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eabf562-d289-4685-8ee5-ed1525930d19" containerName="collect-profiles" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.118160 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.118354 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.121636 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-645jr"] Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.122848 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.166962 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.167210 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.667190133 +0000 UTC m=+146.697161134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.270312 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-catalog-content\") pod \"redhat-operators-645jr\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.270364 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g5kz\" (UniqueName: \"kubernetes.io/projected/10798203-b391-4a87-98a7-b41db2bbb0e2-kube-api-access-8g5kz\") pod \"redhat-operators-645jr\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.270424 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.270451 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-utilities\") pod \"redhat-operators-645jr\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.278931 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.77890733 +0000 UTC m=+146.808878331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.415420 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.416151 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-catalog-content\") pod \"redhat-operators-645jr\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.416193 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g5kz\" (UniqueName: \"kubernetes.io/projected/10798203-b391-4a87-98a7-b41db2bbb0e2-kube-api-access-8g5kz\") pod \"redhat-operators-645jr\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.416238 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-utilities\") pod \"redhat-operators-645jr\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.416765 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-utilities\") pod \"redhat-operators-645jr\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.416860 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-catalog-content\") pod \"redhat-operators-645jr\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.416983 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:02.916961379 +0000 UTC m=+146.946932370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.427170 4676 patch_prober.go:28] interesting pod/downloads-7954f5f757-6m9lm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.427177 4676 patch_prober.go:28] interesting pod/downloads-7954f5f757-6m9lm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.427222 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6m9lm" podUID="4b605a66-7904-4596-a67f-ea21ef41a24b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.427270 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6m9lm" podUID="4b605a66-7904-4596-a67f-ea21ef41a24b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.466114 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sv8x4"] Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.480396 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.497895 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sv8x4"] Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.498915 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g5kz\" (UniqueName: \"kubernetes.io/projected/10798203-b391-4a87-98a7-b41db2bbb0e2-kube-api-access-8g5kz\") pod \"redhat-operators-645jr\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.503661 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.511762 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:02 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:02 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:02 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.512056 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.520067 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.521280 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-catalog-content\") pod \"redhat-operators-sv8x4\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.521497 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6gg\" (UniqueName: \"kubernetes.io/projected/2247b246-d06f-4211-ab24-ff0ee05953b9-kube-api-access-9n6gg\") pod \"redhat-operators-sv8x4\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.521587 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-utilities\") pod \"redhat-operators-sv8x4\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.533867 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.021961551 +0000 UTC m=+147.051932552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.627224 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.627561 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n6gg\" (UniqueName: \"kubernetes.io/projected/2247b246-d06f-4211-ab24-ff0ee05953b9-kube-api-access-9n6gg\") pod \"redhat-operators-sv8x4\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.627591 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-utilities\") pod \"redhat-operators-sv8x4\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.627674 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-catalog-content\") pod \"redhat-operators-sv8x4\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.628900 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.128879115 +0000 UTC m=+147.158850116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.629922 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-utilities\") pod \"redhat-operators-sv8x4\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.630048 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-catalog-content\") pod \"redhat-operators-sv8x4\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.684077 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n6gg\" (UniqueName: \"kubernetes.io/projected/2247b246-d06f-4211-ab24-ff0ee05953b9-kube-api-access-9n6gg\") pod \"redhat-operators-sv8x4\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.684681 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.730949 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.732960 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.232942907 +0000 UTC m=+147.262913908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.760950 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.833968 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.834387 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.334353563 +0000 UTC m=+147.364324564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.840961 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.926001 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7nsm"] Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.935202 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:02 crc kubenswrapper[4676]: E0124 00:06:02.935509 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.435496081 +0000 UTC m=+147.465467082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.944527 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-r6r59" Jan 24 00:06:02 crc kubenswrapper[4676]: I0124 00:06:02.951291 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:02.999282 4676 generic.go:334] "Generic (PLEG): container finished" podID="891b78f7-509c-4e8d-b846-52881396a64d" containerID="5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0" exitCode=0 Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:02.999343 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnqjd" event={"ID":"891b78f7-509c-4e8d-b846-52881396a64d","Type":"ContainerDied","Data":"5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0"} Jan 24 00:06:03 crc kubenswrapper[4676]: W0124 00:06:03.010367 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod920c325a_f36b_4162_9d37_ea88124be938.slice/crio-c30184174638423d11a5d0c24917e087282eea0233650b891df20e30923de797 WatchSource:0}: Error finding container c30184174638423d11a5d0c24917e087282eea0233650b891df20e30923de797: Status 404 returned error can't find the container with id c30184174638423d11a5d0c24917e087282eea0233650b891df20e30923de797 Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.029632 4676 generic.go:334] "Generic (PLEG): container finished" podID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerID="34fdffdd9a497f9a6cd5b90c76fdd621cc1104ee041982343a04412f81b7f452" exitCode=0 Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.029738 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lq6l" event={"ID":"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59","Type":"ContainerDied","Data":"34fdffdd9a497f9a6cd5b90c76fdd621cc1104ee041982343a04412f81b7f452"} Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.036126 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.037222 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.537206068 +0000 UTC m=+147.567177069 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.065555 4676 generic.go:334] "Generic (PLEG): container finished" podID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerID="da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060" exitCode=0 Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.065627 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxmrn" event={"ID":"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd","Type":"ContainerDied","Data":"da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060"} Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.086515 4676 generic.go:334] "Generic (PLEG): container finished" podID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerID="5f55960a6c743513af98906e24b7fac1e853e9cbe06b95180d6ac6353a01cca4" exitCode=0 Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.086591 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxs47" event={"ID":"592859d8-1f7e-4e35-acc9-635e130ad2d2","Type":"ContainerDied","Data":"5f55960a6c743513af98906e24b7fac1e853e9cbe06b95180d6ac6353a01cca4"} Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.121125 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-64phz" event={"ID":"290fd523-8c24-458a-8abb-e32ca43caae1","Type":"ContainerStarted","Data":"5976a48db83ede27ce3462f28b7f1eb3e783f41c258f36eaff6c1057e7c7ed6e"} Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.138275 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.138560 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.638549493 +0000 UTC m=+147.668520484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.151484 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7d9xm" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.154667 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-94fls"] Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.240800 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.241649 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.741635373 +0000 UTC m=+147.771606374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.349441 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.349509 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.349535 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.349589 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.349621 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.350729 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.351603 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.851583154 +0000 UTC m=+147.881554155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.351994 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.355823 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.361607 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.367237 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.369139 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.373329 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.393419 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.415557 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.455412 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.456172 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e53167e0-92bc-428e-909f-791982828360-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e53167e0-92bc-428e-909f-791982828360\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.456245 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e53167e0-92bc-428e-909f-791982828360-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e53167e0-92bc-428e-909f-791982828360\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.456362 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:03.956343307 +0000 UTC m=+147.986314308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.473784 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.487101 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.541842 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:03 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:03 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:03 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.541895 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.557109 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e53167e0-92bc-428e-909f-791982828360-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e53167e0-92bc-428e-909f-791982828360\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.557174 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.557197 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e53167e0-92bc-428e-909f-791982828360-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e53167e0-92bc-428e-909f-791982828360\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.557265 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e53167e0-92bc-428e-909f-791982828360-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e53167e0-92bc-428e-909f-791982828360\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.568184 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.068169447 +0000 UTC m=+148.098140448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.586644 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.658176 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.658391 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.158342215 +0000 UTC m=+148.188313216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.658564 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.658930 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.158922543 +0000 UTC m=+148.188893544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.727551 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e53167e0-92bc-428e-909f-791982828360-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e53167e0-92bc-428e-909f-791982828360\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.759703 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.760125 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.260111313 +0000 UTC m=+148.290082314 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.767338 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.868015 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.868357 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.368344798 +0000 UTC m=+148.398315789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.972721 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:03 crc kubenswrapper[4676]: E0124 00:06:03.976763 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.476728918 +0000 UTC m=+148.506699909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:03 crc kubenswrapper[4676]: I0124 00:06:03.987459 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sv8x4"] Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.040068 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-645jr"] Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.079224 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:04 crc kubenswrapper[4676]: E0124 00:06:04.080031 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.580015415 +0000 UTC m=+148.609986416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.128855 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv8x4" event={"ID":"2247b246-d06f-4211-ab24-ff0ee05953b9","Type":"ContainerStarted","Data":"219fafaa737e407cadd362820da7774d28e18cc7cb1ebb7f3b2b92c72a3323f6"} Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.137107 4676 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.156650 4676 generic.go:334] "Generic (PLEG): container finished" podID="920c325a-f36b-4162-9d37-ea88124be938" containerID="1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3" exitCode=0 Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.156730 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7nsm" event={"ID":"920c325a-f36b-4162-9d37-ea88124be938","Type":"ContainerDied","Data":"1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3"} Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.156757 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7nsm" event={"ID":"920c325a-f36b-4162-9d37-ea88124be938","Type":"ContainerStarted","Data":"c30184174638423d11a5d0c24917e087282eea0233650b891df20e30923de797"} Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.173752 4676 generic.go:334] "Generic (PLEG): container finished" podID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerID="42507e3dd923f3d05f90ce7ba013d39b4e6323bf6515e461e6d463d4272af52b" exitCode=0 Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.173819 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94fls" event={"ID":"236fa0ff-93a4-429d-9a2a-b1ae84167818","Type":"ContainerDied","Data":"42507e3dd923f3d05f90ce7ba013d39b4e6323bf6515e461e6d463d4272af52b"} Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.173846 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94fls" event={"ID":"236fa0ff-93a4-429d-9a2a-b1ae84167818","Type":"ContainerStarted","Data":"0a7c742121571c7169e4b719618bc512b51f0e77919d08a6e5c83dac8827150c"} Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.182671 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.182794 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-64phz" event={"ID":"290fd523-8c24-458a-8abb-e32ca43caae1","Type":"ContainerStarted","Data":"fe80f86e3558468a8feac4e74dd88fd4b9b17729b357f9881bfebe8dfccbc57b"} Jan 24 00:06:04 crc kubenswrapper[4676]: E0124 00:06:04.183107 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.683089906 +0000 UTC m=+148.713060907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.187335 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645jr" event={"ID":"10798203-b391-4a87-98a7-b41db2bbb0e2","Type":"ContainerStarted","Data":"bc7d200b5a89367114e48ce959705259b571653e291943569fe69329c0b533b4"} Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.220859 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-64phz" podStartSLOduration=14.220840364 podStartE2EDuration="14.220840364s" podCreationTimestamp="2026-01-24 00:05:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:06:04.219876983 +0000 UTC m=+148.249847984" watchObservedRunningTime="2026-01-24 00:06:04.220840364 +0000 UTC m=+148.250811365" Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.284045 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:04 crc kubenswrapper[4676]: E0124 00:06:04.285124 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.785107791 +0000 UTC m=+148.815078792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.385911 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:04 crc kubenswrapper[4676]: E0124 00:06:04.386159 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.886133446 +0000 UTC m=+148.916104437 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.386222 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:04 crc kubenswrapper[4676]: E0124 00:06:04.387818 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.887806369 +0000 UTC m=+148.917777370 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shf2w" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:04 crc kubenswrapper[4676]: W0124 00:06:04.392972 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-e3471f8e23142752704577de4b403d882ad5cc2e8b6c5bf1587b44ba83fe9f37 WatchSource:0}: Error finding container e3471f8e23142752704577de4b403d882ad5cc2e8b6c5bf1587b44ba83fe9f37: Status 404 returned error can't find the container with id e3471f8e23142752704577de4b403d882ad5cc2e8b6c5bf1587b44ba83fe9f37 Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.442492 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.488500 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:04 crc kubenswrapper[4676]: E0124 00:06:04.488885 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-24 00:06:04.988858685 +0000 UTC m=+149.018829686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.506537 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:04 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:04 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:04 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.506592 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.556464 4676 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-24T00:06:04.137123834Z","Handler":null,"Name":""} Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.563783 4676 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.563819 4676 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.589937 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.593009 4676 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.593099 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.647093 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shf2w\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.690840 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.701317 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 24 00:06:04 crc kubenswrapper[4676]: I0124 00:06:04.885104 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.232611 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ef7d083c65ea8b6e45bcdfacca6ed93018a0a7b0ce7593b2d127709724c935d3"} Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.232959 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e3471f8e23142752704577de4b403d882ad5cc2e8b6c5bf1587b44ba83fe9f37"} Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.236335 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"17ce290f16c8b20f4eb8341921bdccd3ab48faa94bae57acbcc314e00d5645d0"} Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.236395 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"25b7490cc0d4c59c8fed18314427fc7bd29a53f0fe5fd026e9986cf01bd36a01"} Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.236998 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.242310 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"01e5a5c78cb190c15b5595d924d7ab5c035f776b3eda219cd3cc18aec5fe639b"} Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.242342 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"09b6c0db316c15ca6eabfa1fff99cc04fb02575f7e12598ba07ac53d6dc089a6"} Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.266858 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e53167e0-92bc-428e-909f-791982828360","Type":"ContainerStarted","Data":"484b979f0f3ab8054c80988065d10b132af0af2875a09fa5f2608669bbdb9600"} Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.302857 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shf2w"] Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.328663 4676 generic.go:334] "Generic (PLEG): container finished" podID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerID="eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0" exitCode=0 Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.328819 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645jr" event={"ID":"10798203-b391-4a87-98a7-b41db2bbb0e2","Type":"ContainerDied","Data":"eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0"} Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.376870 4676 generic.go:334] "Generic (PLEG): container finished" podID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerID="07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf" exitCode=0 Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.377136 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv8x4" event={"ID":"2247b246-d06f-4211-ab24-ff0ee05953b9","Type":"ContainerDied","Data":"07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf"} Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.427795 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.427777836 podStartE2EDuration="2.427777836s" podCreationTimestamp="2026-01-24 00:06:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:06:05.421774934 +0000 UTC m=+149.451745935" watchObservedRunningTime="2026-01-24 00:06:05.427777836 +0000 UTC m=+149.457748837" Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.511897 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:05 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:05 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:05 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:05 crc kubenswrapper[4676]: I0124 00:06:05.511946 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.056976 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.058044 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.060401 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.060578 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.069709 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.135227 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7b42804-8738-47a8-8317-33db0f1c4c5b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d7b42804-8738-47a8-8317-33db0f1c4c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.135357 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d7b42804-8738-47a8-8317-33db0f1c4c5b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d7b42804-8738-47a8-8317-33db0f1c4c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.238006 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d7b42804-8738-47a8-8317-33db0f1c4c5b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d7b42804-8738-47a8-8317-33db0f1c4c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.238203 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7b42804-8738-47a8-8317-33db0f1c4c5b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d7b42804-8738-47a8-8317-33db0f1c4c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.239726 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d7b42804-8738-47a8-8317-33db0f1c4c5b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d7b42804-8738-47a8-8317-33db0f1c4c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.284916 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7b42804-8738-47a8-8317-33db0f1c4c5b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d7b42804-8738-47a8-8317-33db0f1c4c5b\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.294102 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.384202 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.445680 4676 generic.go:334] "Generic (PLEG): container finished" podID="e53167e0-92bc-428e-909f-791982828360" containerID="4207fe9c76ca9f8d6b76117b495313275408d130693fe5dcdc925692278cee18" exitCode=0 Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.445955 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e53167e0-92bc-428e-909f-791982828360","Type":"ContainerDied","Data":"4207fe9c76ca9f8d6b76117b495313275408d130693fe5dcdc925692278cee18"} Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.483996 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" event={"ID":"9887557b-81eb-4651-8da2-fd34d7b0be97","Type":"ContainerStarted","Data":"38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472"} Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.484035 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" event={"ID":"9887557b-81eb-4651-8da2-fd34d7b0be97","Type":"ContainerStarted","Data":"927b1ecf53efd6bf7705450ed6ba3d508cc6d33832a7f941121086cc971d4adb"} Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.484066 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.515163 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:06 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:06 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:06 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.515236 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.542488 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" podStartSLOduration=131.542463675 podStartE2EDuration="2m11.542463675s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:06:06.542139484 +0000 UTC m=+150.572110485" watchObservedRunningTime="2026-01-24 00:06:06.542463675 +0000 UTC m=+150.572434676" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.877344 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:06:06 crc kubenswrapper[4676]: I0124 00:06:06.889178 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2scbc" Jan 24 00:06:07 crc kubenswrapper[4676]: I0124 00:06:07.210744 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 24 00:06:07 crc kubenswrapper[4676]: I0124 00:06:07.503612 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d7b42804-8738-47a8-8317-33db0f1c4c5b","Type":"ContainerStarted","Data":"5c407d9f5e5c43a714bcc05894b43ec0b73bbd457e184c3f628e1f23c2cae2eb"} Jan 24 00:06:07 crc kubenswrapper[4676]: I0124 00:06:07.510297 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:07 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:07 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:07 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:07 crc kubenswrapper[4676]: I0124 00:06:07.510577 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:07 crc kubenswrapper[4676]: I0124 00:06:07.941863 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.007948 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e53167e0-92bc-428e-909f-791982828360-kubelet-dir\") pod \"e53167e0-92bc-428e-909f-791982828360\" (UID: \"e53167e0-92bc-428e-909f-791982828360\") " Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.008061 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e53167e0-92bc-428e-909f-791982828360-kube-api-access\") pod \"e53167e0-92bc-428e-909f-791982828360\" (UID: \"e53167e0-92bc-428e-909f-791982828360\") " Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.009450 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e53167e0-92bc-428e-909f-791982828360-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e53167e0-92bc-428e-909f-791982828360" (UID: "e53167e0-92bc-428e-909f-791982828360"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.035620 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e53167e0-92bc-428e-909f-791982828360-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e53167e0-92bc-428e-909f-791982828360" (UID: "e53167e0-92bc-428e-909f-791982828360"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.110007 4676 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e53167e0-92bc-428e-909f-791982828360-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.110036 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e53167e0-92bc-428e-909f-791982828360-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.514775 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:08 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:08 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:08 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.514831 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.549251 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.549571 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e53167e0-92bc-428e-909f-791982828360","Type":"ContainerDied","Data":"484b979f0f3ab8054c80988065d10b132af0af2875a09fa5f2608669bbdb9600"} Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.549588 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="484b979f0f3ab8054c80988065d10b132af0af2875a09fa5f2608669bbdb9600" Jan 24 00:06:08 crc kubenswrapper[4676]: I0124 00:06:08.553893 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-w7sr4" Jan 24 00:06:09 crc kubenswrapper[4676]: I0124 00:06:09.364801 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:06:09 crc kubenswrapper[4676]: I0124 00:06:09.365104 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:06:09 crc kubenswrapper[4676]: I0124 00:06:09.504493 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:09 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:09 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:09 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:09 crc kubenswrapper[4676]: I0124 00:06:09.504547 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:09 crc kubenswrapper[4676]: I0124 00:06:09.601573 4676 generic.go:334] "Generic (PLEG): container finished" podID="d7b42804-8738-47a8-8317-33db0f1c4c5b" containerID="a583295ccb19455325e1ed08ce62db1ebcd517770407d19b13bdbe4a47114fca" exitCode=0 Jan 24 00:06:09 crc kubenswrapper[4676]: I0124 00:06:09.601621 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d7b42804-8738-47a8-8317-33db0f1c4c5b","Type":"ContainerDied","Data":"a583295ccb19455325e1ed08ce62db1ebcd517770407d19b13bdbe4a47114fca"} Jan 24 00:06:10 crc kubenswrapper[4676]: I0124 00:06:10.503928 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:10 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:10 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:10 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:10 crc kubenswrapper[4676]: I0124 00:06:10.504217 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.503139 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:11 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:11 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:11 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.503183 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.582556 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.662852 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d7b42804-8738-47a8-8317-33db0f1c4c5b","Type":"ContainerDied","Data":"5c407d9f5e5c43a714bcc05894b43ec0b73bbd457e184c3f628e1f23c2cae2eb"} Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.662891 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c407d9f5e5c43a714bcc05894b43ec0b73bbd457e184c3f628e1f23c2cae2eb" Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.663159 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.691451 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7b42804-8738-47a8-8317-33db0f1c4c5b-kube-api-access\") pod \"d7b42804-8738-47a8-8317-33db0f1c4c5b\" (UID: \"d7b42804-8738-47a8-8317-33db0f1c4c5b\") " Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.691508 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d7b42804-8738-47a8-8317-33db0f1c4c5b-kubelet-dir\") pod \"d7b42804-8738-47a8-8317-33db0f1c4c5b\" (UID: \"d7b42804-8738-47a8-8317-33db0f1c4c5b\") " Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.691784 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7b42804-8738-47a8-8317-33db0f1c4c5b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d7b42804-8738-47a8-8317-33db0f1c4c5b" (UID: "d7b42804-8738-47a8-8317-33db0f1c4c5b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.714331 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b42804-8738-47a8-8317-33db0f1c4c5b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d7b42804-8738-47a8-8317-33db0f1c4c5b" (UID: "d7b42804-8738-47a8-8317-33db0f1c4c5b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.792919 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7b42804-8738-47a8-8317-33db0f1c4c5b-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:11 crc kubenswrapper[4676]: I0124 00:06:11.792953 4676 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d7b42804-8738-47a8-8317-33db0f1c4c5b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:12 crc kubenswrapper[4676]: I0124 00:06:12.030091 4676 patch_prober.go:28] interesting pod/console-f9d7485db-g2smk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 24 00:06:12 crc kubenswrapper[4676]: I0124 00:06:12.030141 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-g2smk" podUID="5cce043a-2f1b-4f48-967e-c48a00cfe1a6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 24 00:06:12 crc kubenswrapper[4676]: I0124 00:06:12.445876 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6m9lm" Jan 24 00:06:12 crc kubenswrapper[4676]: I0124 00:06:12.505316 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:12 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:12 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:12 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:12 crc kubenswrapper[4676]: I0124 00:06:12.505384 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:13 crc kubenswrapper[4676]: I0124 00:06:13.506051 4676 patch_prober.go:28] interesting pod/router-default-5444994796-zq745 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 24 00:06:13 crc kubenswrapper[4676]: [-]has-synced failed: reason withheld Jan 24 00:06:13 crc kubenswrapper[4676]: [+]process-running ok Jan 24 00:06:13 crc kubenswrapper[4676]: healthz check failed Jan 24 00:06:13 crc kubenswrapper[4676]: I0124 00:06:13.506179 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zq745" podUID="7b8f105b-569d-47f2-b564-a0830b010e31" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 24 00:06:14 crc kubenswrapper[4676]: I0124 00:06:14.505770 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:06:14 crc kubenswrapper[4676]: I0124 00:06:14.508521 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-zq745" Jan 24 00:06:18 crc kubenswrapper[4676]: I0124 00:06:18.514158 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:06:18 crc kubenswrapper[4676]: I0124 00:06:18.533025 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18335446-e572-4741-ad9e-e7aadee7550b-metrics-certs\") pod \"network-metrics-daemon-r4q22\" (UID: \"18335446-e572-4741-ad9e-e7aadee7550b\") " pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:06:18 crc kubenswrapper[4676]: I0124 00:06:18.779620 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-r4q22" Jan 24 00:06:22 crc kubenswrapper[4676]: I0124 00:06:22.035360 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:06:22 crc kubenswrapper[4676]: I0124 00:06:22.042036 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:06:24 crc kubenswrapper[4676]: I0124 00:06:24.891873 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:06:27 crc kubenswrapper[4676]: I0124 00:06:27.890777 4676 generic.go:334] "Generic (PLEG): container finished" podID="00beace7-1e83-40ed-8d92-6da0cae7817f" containerID="e742b4dbf395eb8d16eb17141075162ea559f552eabe8302d3372acaca4a7557" exitCode=0 Jan 24 00:06:27 crc kubenswrapper[4676]: I0124 00:06:27.890918 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29486880-rhzd2" event={"ID":"00beace7-1e83-40ed-8d92-6da0cae7817f","Type":"ContainerDied","Data":"e742b4dbf395eb8d16eb17141075162ea559f552eabe8302d3372acaca4a7557"} Jan 24 00:06:32 crc kubenswrapper[4676]: I0124 00:06:32.312194 4676 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod1eabf562-d289-4685-8ee5-ed1525930d19"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod1eabf562-d289-4685-8ee5-ed1525930d19] : Timed out while waiting for systemd to remove kubepods-burstable-pod1eabf562_d289_4685_8ee5_ed1525930d19.slice" Jan 24 00:06:32 crc kubenswrapper[4676]: E0124 00:06:32.314078 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable pod1eabf562-d289-4685-8ee5-ed1525930d19] : unable to destroy cgroup paths for cgroup [kubepods burstable pod1eabf562-d289-4685-8ee5-ed1525930d19] : Timed out while waiting for systemd to remove kubepods-burstable-pod1eabf562_d289_4685_8ee5_ed1525930d19.slice" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" podUID="1eabf562-d289-4685-8ee5-ed1525930d19" Jan 24 00:06:32 crc kubenswrapper[4676]: I0124 00:06:32.498549 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-ljbnb" Jan 24 00:06:32 crc kubenswrapper[4676]: I0124 00:06:32.936937 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg" Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.364229 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.364631 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.468616 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.512498 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/00beace7-1e83-40ed-8d92-6da0cae7817f-serviceca\") pod \"00beace7-1e83-40ed-8d92-6da0cae7817f\" (UID: \"00beace7-1e83-40ed-8d92-6da0cae7817f\") " Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.512601 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qp87\" (UniqueName: \"kubernetes.io/projected/00beace7-1e83-40ed-8d92-6da0cae7817f-kube-api-access-4qp87\") pod \"00beace7-1e83-40ed-8d92-6da0cae7817f\" (UID: \"00beace7-1e83-40ed-8d92-6da0cae7817f\") " Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.513136 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00beace7-1e83-40ed-8d92-6da0cae7817f-serviceca" (OuterVolumeSpecName: "serviceca") pod "00beace7-1e83-40ed-8d92-6da0cae7817f" (UID: "00beace7-1e83-40ed-8d92-6da0cae7817f"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.524033 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00beace7-1e83-40ed-8d92-6da0cae7817f-kube-api-access-4qp87" (OuterVolumeSpecName: "kube-api-access-4qp87") pod "00beace7-1e83-40ed-8d92-6da0cae7817f" (UID: "00beace7-1e83-40ed-8d92-6da0cae7817f"). InnerVolumeSpecName "kube-api-access-4qp87". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.614851 4676 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/00beace7-1e83-40ed-8d92-6da0cae7817f-serviceca\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.614903 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qp87\" (UniqueName: \"kubernetes.io/projected/00beace7-1e83-40ed-8d92-6da0cae7817f-kube-api-access-4qp87\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.979636 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29486880-rhzd2" event={"ID":"00beace7-1e83-40ed-8d92-6da0cae7817f","Type":"ContainerDied","Data":"79a71111cca31e12c0d06b9a6d103743df145b9a3918dbe29c65faddc57798a0"} Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.979705 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79a71111cca31e12c0d06b9a6d103743df145b9a3918dbe29c65faddc57798a0" Jan 24 00:06:39 crc kubenswrapper[4676]: I0124 00:06:39.979772 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29486880-rhzd2" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.736322 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 24 00:06:42 crc kubenswrapper[4676]: E0124 00:06:42.736922 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b42804-8738-47a8-8317-33db0f1c4c5b" containerName="pruner" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.736934 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b42804-8738-47a8-8317-33db0f1c4c5b" containerName="pruner" Jan 24 00:06:42 crc kubenswrapper[4676]: E0124 00:06:42.736953 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e53167e0-92bc-428e-909f-791982828360" containerName="pruner" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.736959 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="e53167e0-92bc-428e-909f-791982828360" containerName="pruner" Jan 24 00:06:42 crc kubenswrapper[4676]: E0124 00:06:42.736968 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00beace7-1e83-40ed-8d92-6da0cae7817f" containerName="image-pruner" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.736974 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="00beace7-1e83-40ed-8d92-6da0cae7817f" containerName="image-pruner" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.737063 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b42804-8738-47a8-8317-33db0f1c4c5b" containerName="pruner" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.737073 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="00beace7-1e83-40ed-8d92-6da0cae7817f" containerName="image-pruner" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.737084 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="e53167e0-92bc-428e-909f-791982828360" containerName="pruner" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.737497 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.742562 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.742791 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.754438 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b267c5b-c180-4b96-9223-65ed17bd8546-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0b267c5b-c180-4b96-9223-65ed17bd8546\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.754495 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b267c5b-c180-4b96-9223-65ed17bd8546-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0b267c5b-c180-4b96-9223-65ed17bd8546\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.755880 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.855650 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b267c5b-c180-4b96-9223-65ed17bd8546-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0b267c5b-c180-4b96-9223-65ed17bd8546\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.855742 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b267c5b-c180-4b96-9223-65ed17bd8546-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0b267c5b-c180-4b96-9223-65ed17bd8546\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.855845 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b267c5b-c180-4b96-9223-65ed17bd8546-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0b267c5b-c180-4b96-9223-65ed17bd8546\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:42 crc kubenswrapper[4676]: I0124 00:06:42.892907 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b267c5b-c180-4b96-9223-65ed17bd8546-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0b267c5b-c180-4b96-9223-65ed17bd8546\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:43 crc kubenswrapper[4676]: I0124 00:06:43.067516 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:43 crc kubenswrapper[4676]: I0124 00:06:43.491231 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 24 00:06:47 crc kubenswrapper[4676]: E0124 00:06:47.993834 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 24 00:06:47 crc kubenswrapper[4676]: E0124 00:06:47.995194 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8g5kz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-645jr_openshift-marketplace(10798203-b391-4a87-98a7-b41db2bbb0e2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 24 00:06:47 crc kubenswrapper[4676]: E0124 00:06:47.997305 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-645jr" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.147479 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.149335 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.153920 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.250575 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-var-lock\") pod \"installer-9-crc\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.251172 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.251205 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c280bb7c-e926-46c9-bb2c-d88c64634673-kube-api-access\") pod \"installer-9-crc\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.352706 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-var-lock\") pod \"installer-9-crc\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.352783 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.352809 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c280bb7c-e926-46c9-bb2c-d88c64634673-kube-api-access\") pod \"installer-9-crc\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.353319 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-var-lock\") pod \"installer-9-crc\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.353370 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.385831 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c280bb7c-e926-46c9-bb2c-d88c64634673-kube-api-access\") pod \"installer-9-crc\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:48 crc kubenswrapper[4676]: I0124 00:06:48.496655 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:06:49 crc kubenswrapper[4676]: E0124 00:06:49.738940 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-645jr" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" Jan 24 00:06:49 crc kubenswrapper[4676]: E0124 00:06:49.800422 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 24 00:06:49 crc kubenswrapper[4676]: E0124 00:06:49.800592 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n4p5l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-l7nsm_openshift-marketplace(920c325a-f36b-4162-9d37-ea88124be938): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 24 00:06:49 crc kubenswrapper[4676]: E0124 00:06:49.801858 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-l7nsm" podUID="920c325a-f36b-4162-9d37-ea88124be938" Jan 24 00:06:51 crc kubenswrapper[4676]: I0124 00:06:51.966804 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kvcv8"] Jan 24 00:06:52 crc kubenswrapper[4676]: E0124 00:06:52.086408 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-l7nsm" podUID="920c325a-f36b-4162-9d37-ea88124be938" Jan 24 00:06:52 crc kubenswrapper[4676]: E0124 00:06:52.400432 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 24 00:06:52 crc kubenswrapper[4676]: E0124 00:06:52.400570 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jdl5n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9lq6l_openshift-marketplace(53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 24 00:06:52 crc kubenswrapper[4676]: E0124 00:06:52.401694 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9lq6l" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" Jan 24 00:06:53 crc kubenswrapper[4676]: E0124 00:06:53.879782 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 24 00:06:53 crc kubenswrapper[4676]: E0124 00:06:53.879925 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tjhrt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-lxs47_openshift-marketplace(592859d8-1f7e-4e35-acc9-635e130ad2d2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 24 00:06:53 crc kubenswrapper[4676]: E0124 00:06:53.881351 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-lxs47" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" Jan 24 00:06:54 crc kubenswrapper[4676]: E0124 00:06:54.260013 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-lxs47" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" Jan 24 00:06:54 crc kubenswrapper[4676]: E0124 00:06:54.260013 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9lq6l" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" Jan 24 00:06:54 crc kubenswrapper[4676]: I0124 00:06:54.607591 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 24 00:06:54 crc kubenswrapper[4676]: I0124 00:06:54.702200 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 24 00:06:54 crc kubenswrapper[4676]: E0124 00:06:54.726642 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 24 00:06:54 crc kubenswrapper[4676]: E0124 00:06:54.726886 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9n6gg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-sv8x4_openshift-marketplace(2247b246-d06f-4211-ab24-ff0ee05953b9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 24 00:06:54 crc kubenswrapper[4676]: E0124 00:06:54.729191 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-sv8x4" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" Jan 24 00:06:54 crc kubenswrapper[4676]: I0124 00:06:54.762060 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-r4q22"] Jan 24 00:06:54 crc kubenswrapper[4676]: W0124 00:06:54.770188 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18335446_e572_4741_ad9e_e7aadee7550b.slice/crio-52d8453819d216d69f828d5a62928795685c48cac8a247f0f225c6413be7140d WatchSource:0}: Error finding container 52d8453819d216d69f828d5a62928795685c48cac8a247f0f225c6413be7140d: Status 404 returned error can't find the container with id 52d8453819d216d69f828d5a62928795685c48cac8a247f0f225c6413be7140d Jan 24 00:06:55 crc kubenswrapper[4676]: I0124 00:06:55.056753 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0b267c5b-c180-4b96-9223-65ed17bd8546","Type":"ContainerStarted","Data":"988abc66366e3cc44883733bf2fcb7439323f03e26b375f0822d70001d98da5a"} Jan 24 00:06:55 crc kubenswrapper[4676]: I0124 00:06:55.057769 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c280bb7c-e926-46c9-bb2c-d88c64634673","Type":"ContainerStarted","Data":"4ecf5bf658471ce1fe90cd7189121058d1fec23a97202699de0ab189c854a988"} Jan 24 00:06:55 crc kubenswrapper[4676]: I0124 00:06:55.059132 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-r4q22" event={"ID":"18335446-e572-4741-ad9e-e7aadee7550b","Type":"ContainerStarted","Data":"52d8453819d216d69f828d5a62928795685c48cac8a247f0f225c6413be7140d"} Jan 24 00:06:55 crc kubenswrapper[4676]: E0124 00:06:55.060921 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-sv8x4" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" Jan 24 00:06:56 crc kubenswrapper[4676]: I0124 00:06:56.064302 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0b267c5b-c180-4b96-9223-65ed17bd8546","Type":"ContainerStarted","Data":"5108dd50bdc2072552bf66414bb7978d357b797d9814d722f14cc8bcc9f79471"} Jan 24 00:06:56 crc kubenswrapper[4676]: E0124 00:06:56.965498 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 24 00:06:56 crc kubenswrapper[4676]: E0124 00:06:56.965803 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hzqlz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rxmrn_openshift-marketplace(0a4d7c63-cff0-4408-9cb6-450f3ebc53dd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 24 00:06:56 crc kubenswrapper[4676]: E0124 00:06:56.967154 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rxmrn" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" Jan 24 00:06:57 crc kubenswrapper[4676]: I0124 00:06:57.070588 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-r4q22" event={"ID":"18335446-e572-4741-ad9e-e7aadee7550b","Type":"ContainerStarted","Data":"67d0b1beee51807128af4787316e6197966ad653b2c5fbe81128eda05bb4a02c"} Jan 24 00:06:57 crc kubenswrapper[4676]: I0124 00:06:57.073045 4676 generic.go:334] "Generic (PLEG): container finished" podID="0b267c5b-c180-4b96-9223-65ed17bd8546" containerID="5108dd50bdc2072552bf66414bb7978d357b797d9814d722f14cc8bcc9f79471" exitCode=0 Jan 24 00:06:57 crc kubenswrapper[4676]: I0124 00:06:57.073106 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0b267c5b-c180-4b96-9223-65ed17bd8546","Type":"ContainerDied","Data":"5108dd50bdc2072552bf66414bb7978d357b797d9814d722f14cc8bcc9f79471"} Jan 24 00:06:57 crc kubenswrapper[4676]: I0124 00:06:57.077992 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c280bb7c-e926-46c9-bb2c-d88c64634673","Type":"ContainerStarted","Data":"f6bdacd4b420bb177446b935f1e3846bc08e4a95c458b2ca6f67dc97f5745d33"} Jan 24 00:06:57 crc kubenswrapper[4676]: E0124 00:06:57.080018 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rxmrn" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" Jan 24 00:06:57 crc kubenswrapper[4676]: I0124 00:06:57.142448 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=9.142419904 podStartE2EDuration="9.142419904s" podCreationTimestamp="2026-01-24 00:06:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:06:57.127301425 +0000 UTC m=+201.157272436" watchObservedRunningTime="2026-01-24 00:06:57.142419904 +0000 UTC m=+201.172390895" Jan 24 00:06:58 crc kubenswrapper[4676]: I0124 00:06:58.086542 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-r4q22" event={"ID":"18335446-e572-4741-ad9e-e7aadee7550b","Type":"ContainerStarted","Data":"c401c2b5d5d45061d4021b54cdd393ea02dc63cf8afa174d9ad88d789864ac4d"} Jan 24 00:06:58 crc kubenswrapper[4676]: I0124 00:06:58.108892 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-r4q22" podStartSLOduration=183.108874575 podStartE2EDuration="3m3.108874575s" podCreationTimestamp="2026-01-24 00:03:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:06:58.104445267 +0000 UTC m=+202.134416278" watchObservedRunningTime="2026-01-24 00:06:58.108874575 +0000 UTC m=+202.138845576" Jan 24 00:06:58 crc kubenswrapper[4676]: E0124 00:06:58.203704 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 24 00:06:58 crc kubenswrapper[4676]: E0124 00:06:58.203911 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5djr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-94fls_openshift-marketplace(236fa0ff-93a4-429d-9a2a-b1ae84167818): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 24 00:06:58 crc kubenswrapper[4676]: E0124 00:06:58.205124 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-94fls" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" Jan 24 00:06:58 crc kubenswrapper[4676]: I0124 00:06:58.366710 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:58 crc kubenswrapper[4676]: I0124 00:06:58.490531 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b267c5b-c180-4b96-9223-65ed17bd8546-kube-api-access\") pod \"0b267c5b-c180-4b96-9223-65ed17bd8546\" (UID: \"0b267c5b-c180-4b96-9223-65ed17bd8546\") " Jan 24 00:06:58 crc kubenswrapper[4676]: I0124 00:06:58.490803 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b267c5b-c180-4b96-9223-65ed17bd8546-kubelet-dir\") pod \"0b267c5b-c180-4b96-9223-65ed17bd8546\" (UID: \"0b267c5b-c180-4b96-9223-65ed17bd8546\") " Jan 24 00:06:58 crc kubenswrapper[4676]: I0124 00:06:58.490963 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b267c5b-c180-4b96-9223-65ed17bd8546-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0b267c5b-c180-4b96-9223-65ed17bd8546" (UID: "0b267c5b-c180-4b96-9223-65ed17bd8546"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:06:58 crc kubenswrapper[4676]: I0124 00:06:58.510030 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b267c5b-c180-4b96-9223-65ed17bd8546-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b267c5b-c180-4b96-9223-65ed17bd8546" (UID: "0b267c5b-c180-4b96-9223-65ed17bd8546"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:06:58 crc kubenswrapper[4676]: I0124 00:06:58.591791 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b267c5b-c180-4b96-9223-65ed17bd8546-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:58 crc kubenswrapper[4676]: I0124 00:06:58.591827 4676 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0b267c5b-c180-4b96-9223-65ed17bd8546-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:06:59 crc kubenswrapper[4676]: I0124 00:06:59.095482 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 24 00:06:59 crc kubenswrapper[4676]: I0124 00:06:59.096510 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0b267c5b-c180-4b96-9223-65ed17bd8546","Type":"ContainerDied","Data":"988abc66366e3cc44883733bf2fcb7439323f03e26b375f0822d70001d98da5a"} Jan 24 00:06:59 crc kubenswrapper[4676]: I0124 00:06:59.096633 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="988abc66366e3cc44883733bf2fcb7439323f03e26b375f0822d70001d98da5a" Jan 24 00:06:59 crc kubenswrapper[4676]: E0124 00:06:59.099953 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-94fls" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" Jan 24 00:06:59 crc kubenswrapper[4676]: E0124 00:06:59.704172 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 24 00:06:59 crc kubenswrapper[4676]: E0124 00:06:59.704539 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-49sw6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-pnqjd_openshift-marketplace(891b78f7-509c-4e8d-b846-52881396a64d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 24 00:06:59 crc kubenswrapper[4676]: E0124 00:06:59.707179 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-pnqjd" podUID="891b78f7-509c-4e8d-b846-52881396a64d" Jan 24 00:07:00 crc kubenswrapper[4676]: E0124 00:07:00.100717 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-pnqjd" podUID="891b78f7-509c-4e8d-b846-52881396a64d" Jan 24 00:07:07 crc kubenswrapper[4676]: I0124 00:07:07.149700 4676 generic.go:334] "Generic (PLEG): container finished" podID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerID="664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233" exitCode=0 Jan 24 00:07:07 crc kubenswrapper[4676]: I0124 00:07:07.149797 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645jr" event={"ID":"10798203-b391-4a87-98a7-b41db2bbb0e2","Type":"ContainerDied","Data":"664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233"} Jan 24 00:07:07 crc kubenswrapper[4676]: I0124 00:07:07.156118 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv8x4" event={"ID":"2247b246-d06f-4211-ab24-ff0ee05953b9","Type":"ContainerStarted","Data":"0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3"} Jan 24 00:07:07 crc kubenswrapper[4676]: I0124 00:07:07.162691 4676 generic.go:334] "Generic (PLEG): container finished" podID="920c325a-f36b-4162-9d37-ea88124be938" containerID="c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b" exitCode=0 Jan 24 00:07:07 crc kubenswrapper[4676]: I0124 00:07:07.163037 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7nsm" event={"ID":"920c325a-f36b-4162-9d37-ea88124be938","Type":"ContainerDied","Data":"c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b"} Jan 24 00:07:08 crc kubenswrapper[4676]: I0124 00:07:08.169364 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7nsm" event={"ID":"920c325a-f36b-4162-9d37-ea88124be938","Type":"ContainerStarted","Data":"cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211"} Jan 24 00:07:08 crc kubenswrapper[4676]: I0124 00:07:08.171308 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645jr" event={"ID":"10798203-b391-4a87-98a7-b41db2bbb0e2","Type":"ContainerStarted","Data":"1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449"} Jan 24 00:07:08 crc kubenswrapper[4676]: I0124 00:07:08.172678 4676 generic.go:334] "Generic (PLEG): container finished" podID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerID="0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3" exitCode=0 Jan 24 00:07:08 crc kubenswrapper[4676]: I0124 00:07:08.172719 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv8x4" event={"ID":"2247b246-d06f-4211-ab24-ff0ee05953b9","Type":"ContainerDied","Data":"0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3"} Jan 24 00:07:08 crc kubenswrapper[4676]: I0124 00:07:08.220751 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l7nsm" podStartSLOduration=3.7519719719999998 podStartE2EDuration="1m7.22073667s" podCreationTimestamp="2026-01-24 00:06:01 +0000 UTC" firstStartedPulling="2026-01-24 00:06:04.184075117 +0000 UTC m=+148.214046118" lastFinishedPulling="2026-01-24 00:07:07.652839805 +0000 UTC m=+211.682810816" observedRunningTime="2026-01-24 00:07:08.199105779 +0000 UTC m=+212.229076790" watchObservedRunningTime="2026-01-24 00:07:08.22073667 +0000 UTC m=+212.250707671" Jan 24 00:07:08 crc kubenswrapper[4676]: I0124 00:07:08.222607 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-645jr" podStartSLOduration=3.96772861 podStartE2EDuration="1m6.222601487s" podCreationTimestamp="2026-01-24 00:06:02 +0000 UTC" firstStartedPulling="2026-01-24 00:06:05.360618436 +0000 UTC m=+149.390589437" lastFinishedPulling="2026-01-24 00:07:07.615491273 +0000 UTC m=+211.645462314" observedRunningTime="2026-01-24 00:07:08.21903555 +0000 UTC m=+212.249006551" watchObservedRunningTime="2026-01-24 00:07:08.222601487 +0000 UTC m=+212.252572488" Jan 24 00:07:09 crc kubenswrapper[4676]: I0124 00:07:09.182492 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv8x4" event={"ID":"2247b246-d06f-4211-ab24-ff0ee05953b9","Type":"ContainerStarted","Data":"0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3"} Jan 24 00:07:09 crc kubenswrapper[4676]: I0124 00:07:09.364144 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:07:09 crc kubenswrapper[4676]: I0124 00:07:09.364198 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:07:09 crc kubenswrapper[4676]: I0124 00:07:09.364239 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:07:09 crc kubenswrapper[4676]: I0124 00:07:09.364739 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:07:09 crc kubenswrapper[4676]: I0124 00:07:09.364835 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d" gracePeriod=600 Jan 24 00:07:10 crc kubenswrapper[4676]: I0124 00:07:10.190294 4676 generic.go:334] "Generic (PLEG): container finished" podID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerID="d84cf120443faf814099d3b57961bbb33e517c1b14f12a7cc4e82668a41edcc9" exitCode=0 Jan 24 00:07:10 crc kubenswrapper[4676]: I0124 00:07:10.190383 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxs47" event={"ID":"592859d8-1f7e-4e35-acc9-635e130ad2d2","Type":"ContainerDied","Data":"d84cf120443faf814099d3b57961bbb33e517c1b14f12a7cc4e82668a41edcc9"} Jan 24 00:07:10 crc kubenswrapper[4676]: I0124 00:07:10.195557 4676 generic.go:334] "Generic (PLEG): container finished" podID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerID="29117c1fdb99e96004bdef61fde30628f1781d37232b198d232fe3b6cf30fc53" exitCode=0 Jan 24 00:07:10 crc kubenswrapper[4676]: I0124 00:07:10.195675 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lq6l" event={"ID":"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59","Type":"ContainerDied","Data":"29117c1fdb99e96004bdef61fde30628f1781d37232b198d232fe3b6cf30fc53"} Jan 24 00:07:10 crc kubenswrapper[4676]: I0124 00:07:10.201549 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d" exitCode=0 Jan 24 00:07:10 crc kubenswrapper[4676]: I0124 00:07:10.201588 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d"} Jan 24 00:07:10 crc kubenswrapper[4676]: I0124 00:07:10.223927 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sv8x4" podStartSLOduration=5.003844258 podStartE2EDuration="1m8.223894488s" podCreationTimestamp="2026-01-24 00:06:02 +0000 UTC" firstStartedPulling="2026-01-24 00:06:05.400957037 +0000 UTC m=+149.430928038" lastFinishedPulling="2026-01-24 00:07:08.621007267 +0000 UTC m=+212.650978268" observedRunningTime="2026-01-24 00:07:09.199449159 +0000 UTC m=+213.229420160" watchObservedRunningTime="2026-01-24 00:07:10.223894488 +0000 UTC m=+214.253865489" Jan 24 00:07:11 crc kubenswrapper[4676]: I0124 00:07:11.209881 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lq6l" event={"ID":"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59","Type":"ContainerStarted","Data":"c124752b28c2c30e69265c6be80e79a712fa694ec02324e9b77ad0b5952e7fe3"} Jan 24 00:07:11 crc kubenswrapper[4676]: I0124 00:07:11.213785 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"6a2899a2bd48b07d11a54988ae27807805689a40bb41df1e769ca4da39298f4c"} Jan 24 00:07:11 crc kubenswrapper[4676]: I0124 00:07:11.216447 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxs47" event={"ID":"592859d8-1f7e-4e35-acc9-635e130ad2d2","Type":"ContainerStarted","Data":"c53d6e9162fb162b724098a40f36dec02c2bb75d99128d9311568c19c9c217a4"} Jan 24 00:07:11 crc kubenswrapper[4676]: I0124 00:07:11.235178 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9lq6l" podStartSLOduration=4.535611633 podStartE2EDuration="1m12.235163848s" podCreationTimestamp="2026-01-24 00:05:59 +0000 UTC" firstStartedPulling="2026-01-24 00:06:03.032198238 +0000 UTC m=+147.062169239" lastFinishedPulling="2026-01-24 00:07:10.731750453 +0000 UTC m=+214.761721454" observedRunningTime="2026-01-24 00:07:11.234319548 +0000 UTC m=+215.264290559" watchObservedRunningTime="2026-01-24 00:07:11.235163848 +0000 UTC m=+215.265134849" Jan 24 00:07:11 crc kubenswrapper[4676]: I0124 00:07:11.277820 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lxs47" podStartSLOduration=4.607218315 podStartE2EDuration="1m13.277799638s" podCreationTimestamp="2026-01-24 00:05:58 +0000 UTC" firstStartedPulling="2026-01-24 00:06:01.945548796 +0000 UTC m=+145.975519797" lastFinishedPulling="2026-01-24 00:07:10.616130119 +0000 UTC m=+214.646101120" observedRunningTime="2026-01-24 00:07:11.277500208 +0000 UTC m=+215.307471209" watchObservedRunningTime="2026-01-24 00:07:11.277799638 +0000 UTC m=+215.307770649" Jan 24 00:07:11 crc kubenswrapper[4676]: I0124 00:07:11.506421 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:07:11 crc kubenswrapper[4676]: I0124 00:07:11.506505 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:07:12 crc kubenswrapper[4676]: I0124 00:07:12.125592 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:07:12 crc kubenswrapper[4676]: I0124 00:07:12.275810 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:07:12 crc kubenswrapper[4676]: I0124 00:07:12.762438 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:07:12 crc kubenswrapper[4676]: I0124 00:07:12.762507 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:07:12 crc kubenswrapper[4676]: I0124 00:07:12.841515 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:07:12 crc kubenswrapper[4676]: I0124 00:07:12.841885 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:07:13 crc kubenswrapper[4676]: I0124 00:07:13.815024 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-645jr" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="registry-server" probeResult="failure" output=< Jan 24 00:07:13 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:07:13 crc kubenswrapper[4676]: > Jan 24 00:07:13 crc kubenswrapper[4676]: I0124 00:07:13.890543 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sv8x4" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerName="registry-server" probeResult="failure" output=< Jan 24 00:07:13 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:07:13 crc kubenswrapper[4676]: > Jan 24 00:07:15 crc kubenswrapper[4676]: I0124 00:07:15.240724 4676 generic.go:334] "Generic (PLEG): container finished" podID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerID="932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c" exitCode=0 Jan 24 00:07:15 crc kubenswrapper[4676]: I0124 00:07:15.241136 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxmrn" event={"ID":"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd","Type":"ContainerDied","Data":"932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c"} Jan 24 00:07:15 crc kubenswrapper[4676]: I0124 00:07:15.244993 4676 generic.go:334] "Generic (PLEG): container finished" podID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerID="634fec8b6271b2f8d9c4f7db61cb6eed664e6577bbd44282bae63b416d050fcb" exitCode=0 Jan 24 00:07:15 crc kubenswrapper[4676]: I0124 00:07:15.245033 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94fls" event={"ID":"236fa0ff-93a4-429d-9a2a-b1ae84167818","Type":"ContainerDied","Data":"634fec8b6271b2f8d9c4f7db61cb6eed664e6577bbd44282bae63b416d050fcb"} Jan 24 00:07:16 crc kubenswrapper[4676]: I0124 00:07:16.995271 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" podUID="89435826-645d-48a2-aa3b-f5c42003dcbe" containerName="oauth-openshift" containerID="cri-o://b436a79b5d8ef5e9fda6092960f14ec8555b8c360d001d483efb56a39a863e65" gracePeriod=15 Jan 24 00:07:17 crc kubenswrapper[4676]: I0124 00:07:17.263713 4676 generic.go:334] "Generic (PLEG): container finished" podID="891b78f7-509c-4e8d-b846-52881396a64d" containerID="ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903" exitCode=0 Jan 24 00:07:17 crc kubenswrapper[4676]: I0124 00:07:17.263796 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnqjd" event={"ID":"891b78f7-509c-4e8d-b846-52881396a64d","Type":"ContainerDied","Data":"ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903"} Jan 24 00:07:17 crc kubenswrapper[4676]: I0124 00:07:17.277135 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxmrn" event={"ID":"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd","Type":"ContainerStarted","Data":"eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0"} Jan 24 00:07:17 crc kubenswrapper[4676]: I0124 00:07:17.279413 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94fls" event={"ID":"236fa0ff-93a4-429d-9a2a-b1ae84167818","Type":"ContainerStarted","Data":"f8710f8d49b133576915dd9b46891ac9e00701311c21252958d555d9eb5bf410"} Jan 24 00:07:17 crc kubenswrapper[4676]: I0124 00:07:17.315224 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-94fls" podStartSLOduration=4.369324907 podStartE2EDuration="1m16.315198688s" podCreationTimestamp="2026-01-24 00:06:01 +0000 UTC" firstStartedPulling="2026-01-24 00:06:04.185214503 +0000 UTC m=+148.215185504" lastFinishedPulling="2026-01-24 00:07:16.131088284 +0000 UTC m=+220.161059285" observedRunningTime="2026-01-24 00:07:17.311018888 +0000 UTC m=+221.340989899" watchObservedRunningTime="2026-01-24 00:07:17.315198688 +0000 UTC m=+221.345169709" Jan 24 00:07:17 crc kubenswrapper[4676]: I0124 00:07:17.337455 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rxmrn" podStartSLOduration=5.360416809 podStartE2EDuration="1m18.337428011s" podCreationTimestamp="2026-01-24 00:05:59 +0000 UTC" firstStartedPulling="2026-01-24 00:06:03.079560954 +0000 UTC m=+147.109531955" lastFinishedPulling="2026-01-24 00:07:16.056572146 +0000 UTC m=+220.086543157" observedRunningTime="2026-01-24 00:07:17.333473019 +0000 UTC m=+221.363444030" watchObservedRunningTime="2026-01-24 00:07:17.337428011 +0000 UTC m=+221.367399002" Jan 24 00:07:18 crc kubenswrapper[4676]: I0124 00:07:18.286638 4676 generic.go:334] "Generic (PLEG): container finished" podID="89435826-645d-48a2-aa3b-f5c42003dcbe" containerID="b436a79b5d8ef5e9fda6092960f14ec8555b8c360d001d483efb56a39a863e65" exitCode=0 Jan 24 00:07:18 crc kubenswrapper[4676]: I0124 00:07:18.286783 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" event={"ID":"89435826-645d-48a2-aa3b-f5c42003dcbe","Type":"ContainerDied","Data":"b436a79b5d8ef5e9fda6092960f14ec8555b8c360d001d483efb56a39a863e65"} Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.031068 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071691 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlm94\" (UniqueName: \"kubernetes.io/projected/89435826-645d-48a2-aa3b-f5c42003dcbe-kube-api-access-tlm94\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071747 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-serving-cert\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071780 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-login\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071823 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-service-ca\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071843 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-ocp-branding-template\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071872 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-router-certs\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071893 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-error\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071911 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-session\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071935 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-cliconfig\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071955 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-provider-selection\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.071982 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-idp-0-file-data\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.072020 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-dir\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.072056 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-policies\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.072075 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-trusted-ca-bundle\") pod \"89435826-645d-48a2-aa3b-f5c42003dcbe\" (UID: \"89435826-645d-48a2-aa3b-f5c42003dcbe\") " Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.072396 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.074881 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.075284 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.075908 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.093545 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-79b5c48459-zscbn"] Jan 24 00:07:19 crc kubenswrapper[4676]: E0124 00:07:19.094021 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b267c5b-c180-4b96-9223-65ed17bd8546" containerName="pruner" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.094118 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b267c5b-c180-4b96-9223-65ed17bd8546" containerName="pruner" Jan 24 00:07:19 crc kubenswrapper[4676]: E0124 00:07:19.094206 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89435826-645d-48a2-aa3b-f5c42003dcbe" containerName="oauth-openshift" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.094284 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="89435826-645d-48a2-aa3b-f5c42003dcbe" containerName="oauth-openshift" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.094558 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="89435826-645d-48a2-aa3b-f5c42003dcbe" containerName="oauth-openshift" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.094652 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b267c5b-c180-4b96-9223-65ed17bd8546" containerName="pruner" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.095516 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.094279 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.102065 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79b5c48459-zscbn"] Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.106959 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89435826-645d-48a2-aa3b-f5c42003dcbe-kube-api-access-tlm94" (OuterVolumeSpecName: "kube-api-access-tlm94") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "kube-api-access-tlm94". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.113456 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.114327 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.116004 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.116653 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.124473 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.128472 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.128817 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.129179 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "89435826-645d-48a2-aa3b-f5c42003dcbe" (UID: "89435826-645d-48a2-aa3b-f5c42003dcbe"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.173261 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-template-error\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.173574 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v29c7\" (UniqueName: \"kubernetes.io/projected/66786051-105a-4770-ba4f-25f9b16bc2bb-kube-api-access-v29c7\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.173699 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-service-ca\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.173805 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.173909 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.174012 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-router-certs\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.174305 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-audit-policies\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.174450 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66786051-105a-4770-ba4f-25f9b16bc2bb-audit-dir\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.174573 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.174146 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.174684 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.174740 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.174900 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-template-login\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.175022 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.175362 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.175523 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-session\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.175657 4676 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.175739 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.175805 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlm94\" (UniqueName: \"kubernetes.io/projected/89435826-645d-48a2-aa3b-f5c42003dcbe-kube-api-access-tlm94\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.175881 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.175954 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.176010 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.176066 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.176127 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.176187 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.176261 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.176324 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.176415 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.176483 4676 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/89435826-645d-48a2-aa3b-f5c42003dcbe-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.176542 4676 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/89435826-645d-48a2-aa3b-f5c42003dcbe-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.218347 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278273 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-template-error\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278363 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v29c7\" (UniqueName: \"kubernetes.io/projected/66786051-105a-4770-ba4f-25f9b16bc2bb-kube-api-access-v29c7\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278413 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-service-ca\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278443 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278468 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278492 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-router-certs\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278523 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-audit-policies\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278544 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66786051-105a-4770-ba4f-25f9b16bc2bb-audit-dir\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278564 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278584 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278612 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278636 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-template-login\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278660 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278704 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-session\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.278932 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66786051-105a-4770-ba4f-25f9b16bc2bb-audit-dir\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.281053 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-service-ca\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.281856 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-audit-policies\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.282825 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.282885 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.286009 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.286430 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.286828 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.286944 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-session\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.287711 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-router-certs\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.288078 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-template-error\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.289926 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-user-template-login\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.294355 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66786051-105a-4770-ba4f-25f9b16bc2bb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.298687 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v29c7\" (UniqueName: \"kubernetes.io/projected/66786051-105a-4770-ba4f-25f9b16bc2bb-kube-api-access-v29c7\") pod \"oauth-openshift-79b5c48459-zscbn\" (UID: \"66786051-105a-4770-ba4f-25f9b16bc2bb\") " pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.300164 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" event={"ID":"89435826-645d-48a2-aa3b-f5c42003dcbe","Type":"ContainerDied","Data":"f8acee532eb12a14f5c8f001c0b4e4969213cd1ade4bba8504365d2ac03f9941"} Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.300248 4676 scope.go:117] "RemoveContainer" containerID="b436a79b5d8ef5e9fda6092960f14ec8555b8c360d001d483efb56a39a863e65" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.300755 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-kvcv8" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.344745 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kvcv8"] Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.345182 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-kvcv8"] Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.352540 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.432722 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.432944 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.463055 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.478335 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.720560 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.722318 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.763593 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:07:19 crc kubenswrapper[4676]: I0124 00:07:19.984662 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-79b5c48459-zscbn"] Jan 24 00:07:20 crc kubenswrapper[4676]: I0124 00:07:20.266308 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89435826-645d-48a2-aa3b-f5c42003dcbe" path="/var/lib/kubelet/pods/89435826-645d-48a2-aa3b-f5c42003dcbe/volumes" Jan 24 00:07:20 crc kubenswrapper[4676]: I0124 00:07:20.307809 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" event={"ID":"66786051-105a-4770-ba4f-25f9b16bc2bb","Type":"ContainerStarted","Data":"b7ccf8f1b0c31882e06a07a8fca815211dc5741c6152d88fac1459c3e3197010"} Jan 24 00:07:20 crc kubenswrapper[4676]: I0124 00:07:20.361149 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:07:21 crc kubenswrapper[4676]: I0124 00:07:21.330471 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lq6l"] Jan 24 00:07:21 crc kubenswrapper[4676]: I0124 00:07:21.368599 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:07:21 crc kubenswrapper[4676]: I0124 00:07:21.903276 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:07:21 crc kubenswrapper[4676]: I0124 00:07:21.903640 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:07:21 crc kubenswrapper[4676]: I0124 00:07:21.949958 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:07:22 crc kubenswrapper[4676]: I0124 00:07:22.327864 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnqjd" event={"ID":"891b78f7-509c-4e8d-b846-52881396a64d","Type":"ContainerStarted","Data":"fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461"} Jan 24 00:07:22 crc kubenswrapper[4676]: I0124 00:07:22.329745 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" event={"ID":"66786051-105a-4770-ba4f-25f9b16bc2bb","Type":"ContainerStarted","Data":"528a5f11d379d8a4ae5dde74036af662d2236007e58088a64dec654b7996ed80"} Jan 24 00:07:22 crc kubenswrapper[4676]: I0124 00:07:22.354516 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pnqjd" podStartSLOduration=4.957957099 podStartE2EDuration="1m23.354487407s" podCreationTimestamp="2026-01-24 00:05:59 +0000 UTC" firstStartedPulling="2026-01-24 00:06:03.009726648 +0000 UTC m=+147.039697649" lastFinishedPulling="2026-01-24 00:07:21.406256936 +0000 UTC m=+225.436227957" observedRunningTime="2026-01-24 00:07:22.350421322 +0000 UTC m=+226.380392353" watchObservedRunningTime="2026-01-24 00:07:22.354487407 +0000 UTC m=+226.384458428" Jan 24 00:07:22 crc kubenswrapper[4676]: I0124 00:07:22.396722 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:07:22 crc kubenswrapper[4676]: I0124 00:07:22.419019 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" podStartSLOduration=31.418994948 podStartE2EDuration="31.418994948s" podCreationTimestamp="2026-01-24 00:06:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:07:22.381999208 +0000 UTC m=+226.411970239" watchObservedRunningTime="2026-01-24 00:07:22.418994948 +0000 UTC m=+226.448965989" Jan 24 00:07:22 crc kubenswrapper[4676]: I0124 00:07:22.803921 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:07:22 crc kubenswrapper[4676]: I0124 00:07:22.867763 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:07:22 crc kubenswrapper[4676]: I0124 00:07:22.923975 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:07:22 crc kubenswrapper[4676]: I0124 00:07:22.968586 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:07:23 crc kubenswrapper[4676]: I0124 00:07:23.336121 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9lq6l" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerName="registry-server" containerID="cri-o://c124752b28c2c30e69265c6be80e79a712fa694ec02324e9b77ad0b5952e7fe3" gracePeriod=2 Jan 24 00:07:23 crc kubenswrapper[4676]: I0124 00:07:23.337211 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:23 crc kubenswrapper[4676]: I0124 00:07:23.342406 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-79b5c48459-zscbn" Jan 24 00:07:23 crc kubenswrapper[4676]: I0124 00:07:23.529485 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-94fls"] Jan 24 00:07:24 crc kubenswrapper[4676]: I0124 00:07:24.342625 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-94fls" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerName="registry-server" containerID="cri-o://f8710f8d49b133576915dd9b46891ac9e00701311c21252958d555d9eb5bf410" gracePeriod=2 Jan 24 00:07:25 crc kubenswrapper[4676]: I0124 00:07:25.930218 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sv8x4"] Jan 24 00:07:25 crc kubenswrapper[4676]: I0124 00:07:25.931066 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sv8x4" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerName="registry-server" containerID="cri-o://0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3" gracePeriod=2 Jan 24 00:07:26 crc kubenswrapper[4676]: I0124 00:07:26.360329 4676 generic.go:334] "Generic (PLEG): container finished" podID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerID="c124752b28c2c30e69265c6be80e79a712fa694ec02324e9b77ad0b5952e7fe3" exitCode=0 Jan 24 00:07:26 crc kubenswrapper[4676]: I0124 00:07:26.360424 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lq6l" event={"ID":"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59","Type":"ContainerDied","Data":"c124752b28c2c30e69265c6be80e79a712fa694ec02324e9b77ad0b5952e7fe3"} Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.370066 4676 generic.go:334] "Generic (PLEG): container finished" podID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerID="f8710f8d49b133576915dd9b46891ac9e00701311c21252958d555d9eb5bf410" exitCode=0 Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.370241 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94fls" event={"ID":"236fa0ff-93a4-429d-9a2a-b1ae84167818","Type":"ContainerDied","Data":"f8710f8d49b133576915dd9b46891ac9e00701311c21252958d555d9eb5bf410"} Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.505502 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.603371 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-utilities\") pod \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.604017 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdl5n\" (UniqueName: \"kubernetes.io/projected/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-kube-api-access-jdl5n\") pod \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.604150 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-catalog-content\") pod \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\" (UID: \"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59\") " Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.604595 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-utilities" (OuterVolumeSpecName: "utilities") pod "53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" (UID: "53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.604888 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.615624 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-kube-api-access-jdl5n" (OuterVolumeSpecName: "kube-api-access-jdl5n") pod "53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" (UID: "53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59"). InnerVolumeSpecName "kube-api-access-jdl5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.660980 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" (UID: "53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.707263 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.707303 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdl5n\" (UniqueName: \"kubernetes.io/projected/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59-kube-api-access-jdl5n\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:27 crc kubenswrapper[4676]: I0124 00:07:27.874700 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.010483 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-utilities\") pod \"236fa0ff-93a4-429d-9a2a-b1ae84167818\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.010597 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5djr\" (UniqueName: \"kubernetes.io/projected/236fa0ff-93a4-429d-9a2a-b1ae84167818-kube-api-access-g5djr\") pod \"236fa0ff-93a4-429d-9a2a-b1ae84167818\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.010677 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-catalog-content\") pod \"236fa0ff-93a4-429d-9a2a-b1ae84167818\" (UID: \"236fa0ff-93a4-429d-9a2a-b1ae84167818\") " Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.011254 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-utilities" (OuterVolumeSpecName: "utilities") pod "236fa0ff-93a4-429d-9a2a-b1ae84167818" (UID: "236fa0ff-93a4-429d-9a2a-b1ae84167818"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.013220 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/236fa0ff-93a4-429d-9a2a-b1ae84167818-kube-api-access-g5djr" (OuterVolumeSpecName: "kube-api-access-g5djr") pod "236fa0ff-93a4-429d-9a2a-b1ae84167818" (UID: "236fa0ff-93a4-429d-9a2a-b1ae84167818"). InnerVolumeSpecName "kube-api-access-g5djr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.031934 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "236fa0ff-93a4-429d-9a2a-b1ae84167818" (UID: "236fa0ff-93a4-429d-9a2a-b1ae84167818"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.112336 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5djr\" (UniqueName: \"kubernetes.io/projected/236fa0ff-93a4-429d-9a2a-b1ae84167818-kube-api-access-g5djr\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.112380 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.112400 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/236fa0ff-93a4-429d-9a2a-b1ae84167818-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.346747 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.376689 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lq6l" event={"ID":"53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59","Type":"ContainerDied","Data":"240a15f7e6730a7c6ad0337131f8a5c437c590dd34ae29a62a731e625e9f6839"} Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.376746 4676 scope.go:117] "RemoveContainer" containerID="c124752b28c2c30e69265c6be80e79a712fa694ec02324e9b77ad0b5952e7fe3" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.376891 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lq6l" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.381781 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-94fls" event={"ID":"236fa0ff-93a4-429d-9a2a-b1ae84167818","Type":"ContainerDied","Data":"0a7c742121571c7169e4b719618bc512b51f0e77919d08a6e5c83dac8827150c"} Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.381890 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-94fls" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.383444 4676 generic.go:334] "Generic (PLEG): container finished" podID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerID="0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3" exitCode=0 Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.383471 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv8x4" event={"ID":"2247b246-d06f-4211-ab24-ff0ee05953b9","Type":"ContainerDied","Data":"0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3"} Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.383491 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sv8x4" event={"ID":"2247b246-d06f-4211-ab24-ff0ee05953b9","Type":"ContainerDied","Data":"219fafaa737e407cadd362820da7774d28e18cc7cb1ebb7f3b2b92c72a3323f6"} Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.384429 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sv8x4" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.396859 4676 scope.go:117] "RemoveContainer" containerID="29117c1fdb99e96004bdef61fde30628f1781d37232b198d232fe3b6cf30fc53" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.403948 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lq6l"] Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.403993 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9lq6l"] Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.425644 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-94fls"] Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.426846 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-94fls"] Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.432146 4676 scope.go:117] "RemoveContainer" containerID="34fdffdd9a497f9a6cd5b90c76fdd621cc1104ee041982343a04412f81b7f452" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.453071 4676 scope.go:117] "RemoveContainer" containerID="f8710f8d49b133576915dd9b46891ac9e00701311c21252958d555d9eb5bf410" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.474311 4676 scope.go:117] "RemoveContainer" containerID="634fec8b6271b2f8d9c4f7db61cb6eed664e6577bbd44282bae63b416d050fcb" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.489245 4676 scope.go:117] "RemoveContainer" containerID="42507e3dd923f3d05f90ce7ba013d39b4e6323bf6515e461e6d463d4272af52b" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.505451 4676 scope.go:117] "RemoveContainer" containerID="0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.519677 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n6gg\" (UniqueName: \"kubernetes.io/projected/2247b246-d06f-4211-ab24-ff0ee05953b9-kube-api-access-9n6gg\") pod \"2247b246-d06f-4211-ab24-ff0ee05953b9\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.519736 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-utilities\") pod \"2247b246-d06f-4211-ab24-ff0ee05953b9\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.519861 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-catalog-content\") pod \"2247b246-d06f-4211-ab24-ff0ee05953b9\" (UID: \"2247b246-d06f-4211-ab24-ff0ee05953b9\") " Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.521396 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-utilities" (OuterVolumeSpecName: "utilities") pod "2247b246-d06f-4211-ab24-ff0ee05953b9" (UID: "2247b246-d06f-4211-ab24-ff0ee05953b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.524474 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2247b246-d06f-4211-ab24-ff0ee05953b9-kube-api-access-9n6gg" (OuterVolumeSpecName: "kube-api-access-9n6gg") pod "2247b246-d06f-4211-ab24-ff0ee05953b9" (UID: "2247b246-d06f-4211-ab24-ff0ee05953b9"). InnerVolumeSpecName "kube-api-access-9n6gg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.527712 4676 scope.go:117] "RemoveContainer" containerID="0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.543834 4676 scope.go:117] "RemoveContainer" containerID="07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.560567 4676 scope.go:117] "RemoveContainer" containerID="0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3" Jan 24 00:07:28 crc kubenswrapper[4676]: E0124 00:07:28.561200 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3\": container with ID starting with 0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3 not found: ID does not exist" containerID="0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.561231 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3"} err="failed to get container status \"0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3\": rpc error: code = NotFound desc = could not find container \"0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3\": container with ID starting with 0b195c50824310367359504ebc2c1523df4dda14dc8c6d70328cd706ff0670e3 not found: ID does not exist" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.561253 4676 scope.go:117] "RemoveContainer" containerID="0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3" Jan 24 00:07:28 crc kubenswrapper[4676]: E0124 00:07:28.561766 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3\": container with ID starting with 0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3 not found: ID does not exist" containerID="0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.561822 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3"} err="failed to get container status \"0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3\": rpc error: code = NotFound desc = could not find container \"0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3\": container with ID starting with 0ef99f4ef6f0988afd2cf781f45853cf9ddb0aace684a1d3f7d73355a726b8d3 not found: ID does not exist" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.561860 4676 scope.go:117] "RemoveContainer" containerID="07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf" Jan 24 00:07:28 crc kubenswrapper[4676]: E0124 00:07:28.562275 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf\": container with ID starting with 07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf not found: ID does not exist" containerID="07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.562310 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf"} err="failed to get container status \"07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf\": rpc error: code = NotFound desc = could not find container \"07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf\": container with ID starting with 07acdb0762573bb2b8235a6edcf0cfe4842b7bcb639741acfb88ccb64944ecbf not found: ID does not exist" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.621426 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n6gg\" (UniqueName: \"kubernetes.io/projected/2247b246-d06f-4211-ab24-ff0ee05953b9-kube-api-access-9n6gg\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.621628 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.646085 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2247b246-d06f-4211-ab24-ff0ee05953b9" (UID: "2247b246-d06f-4211-ab24-ff0ee05953b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.712135 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sv8x4"] Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.725233 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2247b246-d06f-4211-ab24-ff0ee05953b9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:28 crc kubenswrapper[4676]: I0124 00:07:28.727982 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sv8x4"] Jan 24 00:07:29 crc kubenswrapper[4676]: I0124 00:07:29.948294 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:07:29 crc kubenswrapper[4676]: I0124 00:07:29.948794 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:07:30 crc kubenswrapper[4676]: I0124 00:07:30.017338 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:07:30 crc kubenswrapper[4676]: I0124 00:07:30.267727 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" path="/var/lib/kubelet/pods/2247b246-d06f-4211-ab24-ff0ee05953b9/volumes" Jan 24 00:07:30 crc kubenswrapper[4676]: I0124 00:07:30.268459 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" path="/var/lib/kubelet/pods/236fa0ff-93a4-429d-9a2a-b1ae84167818/volumes" Jan 24 00:07:30 crc kubenswrapper[4676]: I0124 00:07:30.269111 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" path="/var/lib/kubelet/pods/53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59/volumes" Jan 24 00:07:30 crc kubenswrapper[4676]: I0124 00:07:30.478505 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.746704 4676 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.747286 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerName="extract-content" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747297 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerName="extract-content" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.747305 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerName="extract-utilities" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747310 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerName="extract-utilities" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.747320 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerName="extract-content" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747326 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerName="extract-content" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.747333 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerName="extract-utilities" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747339 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerName="extract-utilities" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.747351 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerName="registry-server" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747357 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerName="registry-server" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.747366 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerName="extract-content" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747371 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerName="extract-content" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.747382 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerName="extract-utilities" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747388 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerName="extract-utilities" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.747409 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerName="registry-server" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747417 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerName="registry-server" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.747427 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerName="registry-server" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747432 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerName="registry-server" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747522 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2247b246-d06f-4211-ab24-ff0ee05953b9" containerName="registry-server" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747535 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="53ec2f7b-a5ba-41ba-9e8b-2a04032f7e59" containerName="registry-server" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747543 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="236fa0ff-93a4-429d-9a2a-b1ae84167818" containerName="registry-server" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747846 4676 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.747967 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748073 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142" gracePeriod=15 Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748137 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa" gracePeriod=15 Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748175 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b" gracePeriod=15 Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748194 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a" gracePeriod=15 Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748137 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9" gracePeriod=15 Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748578 4676 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.748762 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748773 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.748780 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748786 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.748797 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748803 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.748809 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748815 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.748829 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748834 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.748843 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748848 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748931 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748940 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748948 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748958 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.748965 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 24 00:07:33 crc kubenswrapper[4676]: E0124 00:07:33.749061 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.749067 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.749144 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.787492 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.842588 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.842633 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.842653 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.842673 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.842962 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.843018 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.843059 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.843168 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.944809 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.944874 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.944902 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.944952 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.944985 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.944986 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945031 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945069 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945008 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945013 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945043 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945124 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945147 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945069 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945186 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:33 crc kubenswrapper[4676]: I0124 00:07:33.945207 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.085429 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:07:34 crc kubenswrapper[4676]: W0124 00:07:34.101637 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-976837c5d5eaa84109b1d0a099c5a8b73d4b2e0dae46767124489e7eb0947d07 WatchSource:0}: Error finding container 976837c5d5eaa84109b1d0a099c5a8b73d4b2e0dae46767124489e7eb0947d07: Status 404 returned error can't find the container with id 976837c5d5eaa84109b1d0a099c5a8b73d4b2e0dae46767124489e7eb0947d07 Jan 24 00:07:34 crc kubenswrapper[4676]: E0124 00:07:34.104201 4676 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.27:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d8214b6bb6799 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-24 00:07:34.103164825 +0000 UTC m=+238.133135856,LastTimestamp:2026-01-24 00:07:34.103164825 +0000 UTC m=+238.133135856,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.433751 4676 generic.go:334] "Generic (PLEG): container finished" podID="c280bb7c-e926-46c9-bb2c-d88c64634673" containerID="f6bdacd4b420bb177446b935f1e3846bc08e4a95c458b2ca6f67dc97f5745d33" exitCode=0 Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.433815 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c280bb7c-e926-46c9-bb2c-d88c64634673","Type":"ContainerDied","Data":"f6bdacd4b420bb177446b935f1e3846bc08e4a95c458b2ca6f67dc97f5745d33"} Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.434439 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.434817 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.436551 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.437697 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.438336 4676 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a" exitCode=0 Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.438359 4676 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9" exitCode=0 Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.438369 4676 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa" exitCode=0 Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.438394 4676 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b" exitCode=2 Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.438470 4676 scope.go:117] "RemoveContainer" containerID="5c47cad0fa09af5d6054694c41c2fd3ce35fdc093e673df7283b56ad009a05fc" Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.440168 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"31446979da9b2814d80beb1cd72c15277eaaac69a2a2c9676d049a6a8acf10c5"} Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.440201 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"976837c5d5eaa84109b1d0a099c5a8b73d4b2e0dae46767124489e7eb0947d07"} Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.440954 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:34 crc kubenswrapper[4676]: I0124 00:07:34.441247 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:35 crc kubenswrapper[4676]: E0124 00:07:35.296006 4676 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.27:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d8214b6bb6799 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-24 00:07:34.103164825 +0000 UTC m=+238.133135856,LastTimestamp:2026-01-24 00:07:34.103164825 +0000 UTC m=+238.133135856,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.449414 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.697765 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.698713 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.698867 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.766106 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-var-lock\") pod \"c280bb7c-e926-46c9-bb2c-d88c64634673\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.766203 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-kubelet-dir\") pod \"c280bb7c-e926-46c9-bb2c-d88c64634673\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.766236 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c280bb7c-e926-46c9-bb2c-d88c64634673-kube-api-access\") pod \"c280bb7c-e926-46c9-bb2c-d88c64634673\" (UID: \"c280bb7c-e926-46c9-bb2c-d88c64634673\") " Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.766622 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-var-lock" (OuterVolumeSpecName: "var-lock") pod "c280bb7c-e926-46c9-bb2c-d88c64634673" (UID: "c280bb7c-e926-46c9-bb2c-d88c64634673"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.766645 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c280bb7c-e926-46c9-bb2c-d88c64634673" (UID: "c280bb7c-e926-46c9-bb2c-d88c64634673"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.784436 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c280bb7c-e926-46c9-bb2c-d88c64634673-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c280bb7c-e926-46c9-bb2c-d88c64634673" (UID: "c280bb7c-e926-46c9-bb2c-d88c64634673"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.870140 4676 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-var-lock\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.870173 4676 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c280bb7c-e926-46c9-bb2c-d88c64634673-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:35 crc kubenswrapper[4676]: I0124 00:07:35.870186 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c280bb7c-e926-46c9-bb2c-d88c64634673-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.113151 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.114534 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.115883 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.116519 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.117168 4676 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.258287 4676 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.258573 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.258799 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.274142 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.274226 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.274253 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.274327 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.274580 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.274668 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.275115 4676 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.275139 4676 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.275197 4676 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.457397 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c280bb7c-e926-46c9-bb2c-d88c64634673","Type":"ContainerDied","Data":"4ecf5bf658471ce1fe90cd7189121058d1fec23a97202699de0ab189c854a988"} Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.457424 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.457452 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ecf5bf658471ce1fe90cd7189121058d1fec23a97202699de0ab189c854a988" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.460619 4676 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142" exitCode=0 Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.460698 4676 scope.go:117] "RemoveContainer" containerID="f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.460744 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.461551 4676 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.461885 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.463609 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.464023 4676 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.471153 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.471853 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.478410 4676 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.478786 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.479226 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.480842 4676 scope.go:117] "RemoveContainer" containerID="7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.495503 4676 scope.go:117] "RemoveContainer" containerID="229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.522120 4676 scope.go:117] "RemoveContainer" containerID="1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.536982 4676 scope.go:117] "RemoveContainer" containerID="85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.550997 4676 scope.go:117] "RemoveContainer" containerID="53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.581871 4676 scope.go:117] "RemoveContainer" containerID="f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a" Jan 24 00:07:36 crc kubenswrapper[4676]: E0124 00:07:36.582569 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\": container with ID starting with f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a not found: ID does not exist" containerID="f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.582613 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a"} err="failed to get container status \"f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\": rpc error: code = NotFound desc = could not find container \"f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a\": container with ID starting with f16d2dabd09c8f6638d8a22e94054b9edf5285e43e2c92c32684d973cb01f33a not found: ID does not exist" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.582649 4676 scope.go:117] "RemoveContainer" containerID="7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9" Jan 24 00:07:36 crc kubenswrapper[4676]: E0124 00:07:36.583946 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\": container with ID starting with 7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9 not found: ID does not exist" containerID="7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.583978 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9"} err="failed to get container status \"7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\": rpc error: code = NotFound desc = could not find container \"7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9\": container with ID starting with 7d29a6a014f64831d9a51bfe94c8dee076d996d3ae19a14b236d784d365757c9 not found: ID does not exist" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.584003 4676 scope.go:117] "RemoveContainer" containerID="229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa" Jan 24 00:07:36 crc kubenswrapper[4676]: E0124 00:07:36.584244 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\": container with ID starting with 229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa not found: ID does not exist" containerID="229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.584277 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa"} err="failed to get container status \"229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\": rpc error: code = NotFound desc = could not find container \"229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa\": container with ID starting with 229ff7cce0b4e89faae3c3c05ca116c9b91bd68ea3975e434f56778bef20f3aa not found: ID does not exist" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.584300 4676 scope.go:117] "RemoveContainer" containerID="1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b" Jan 24 00:07:36 crc kubenswrapper[4676]: E0124 00:07:36.584648 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\": container with ID starting with 1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b not found: ID does not exist" containerID="1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.584675 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b"} err="failed to get container status \"1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\": rpc error: code = NotFound desc = could not find container \"1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b\": container with ID starting with 1f7445ac3f25b14c2fa8ab85a43fb52bfe9a61373e1631b03ac314701897d57b not found: ID does not exist" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.584691 4676 scope.go:117] "RemoveContainer" containerID="85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142" Jan 24 00:07:36 crc kubenswrapper[4676]: E0124 00:07:36.584956 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\": container with ID starting with 85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142 not found: ID does not exist" containerID="85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.584976 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142"} err="failed to get container status \"85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\": rpc error: code = NotFound desc = could not find container \"85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142\": container with ID starting with 85fef114ba2a21932b6ff9a529a74938a6446ac89355fe1ab4b7d447194c4142 not found: ID does not exist" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.585023 4676 scope.go:117] "RemoveContainer" containerID="53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c" Jan 24 00:07:36 crc kubenswrapper[4676]: E0124 00:07:36.585418 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\": container with ID starting with 53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c not found: ID does not exist" containerID="53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c" Jan 24 00:07:36 crc kubenswrapper[4676]: I0124 00:07:36.585452 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c"} err="failed to get container status \"53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\": rpc error: code = NotFound desc = could not find container \"53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c\": container with ID starting with 53647b8b73c8af266a7c6abbd5af86c45dabbc3ccd4564adfcc4d41952518b4c not found: ID does not exist" Jan 24 00:07:38 crc kubenswrapper[4676]: I0124 00:07:38.266462 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 24 00:07:40 crc kubenswrapper[4676]: E0124 00:07:40.423044 4676 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:40 crc kubenswrapper[4676]: E0124 00:07:40.424086 4676 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:40 crc kubenswrapper[4676]: E0124 00:07:40.424353 4676 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:40 crc kubenswrapper[4676]: E0124 00:07:40.424758 4676 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:40 crc kubenswrapper[4676]: E0124 00:07:40.424923 4676 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:40 crc kubenswrapper[4676]: I0124 00:07:40.424947 4676 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 24 00:07:40 crc kubenswrapper[4676]: E0124 00:07:40.425089 4676 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" interval="200ms" Jan 24 00:07:40 crc kubenswrapper[4676]: E0124 00:07:40.626425 4676 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" interval="400ms" Jan 24 00:07:41 crc kubenswrapper[4676]: E0124 00:07:41.027999 4676 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" interval="800ms" Jan 24 00:07:41 crc kubenswrapper[4676]: E0124 00:07:41.829012 4676 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" interval="1.6s" Jan 24 00:07:42 crc kubenswrapper[4676]: E0124 00:07:42.348982 4676 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.27:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" volumeName="registry-storage" Jan 24 00:07:43 crc kubenswrapper[4676]: E0124 00:07:43.430549 4676 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" interval="3.2s" Jan 24 00:07:45 crc kubenswrapper[4676]: E0124 00:07:45.297357 4676 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.27:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188d8214b6bb6799 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-24 00:07:34.103164825 +0000 UTC m=+238.133135856,LastTimestamp:2026-01-24 00:07:34.103164825 +0000 UTC m=+238.133135856,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 24 00:07:46 crc kubenswrapper[4676]: I0124 00:07:46.257658 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:46 crc kubenswrapper[4676]: I0124 00:07:46.258006 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:46 crc kubenswrapper[4676]: E0124 00:07:46.632046 4676 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.27:6443: connect: connection refused" interval="6.4s" Jan 24 00:07:47 crc kubenswrapper[4676]: I0124 00:07:47.255459 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:47 crc kubenswrapper[4676]: I0124 00:07:47.256768 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:47 crc kubenswrapper[4676]: I0124 00:07:47.257166 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:47 crc kubenswrapper[4676]: I0124 00:07:47.278049 4676 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:47 crc kubenswrapper[4676]: I0124 00:07:47.278109 4676 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:47 crc kubenswrapper[4676]: E0124 00:07:47.278599 4676 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:47 crc kubenswrapper[4676]: I0124 00:07:47.279232 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:47 crc kubenswrapper[4676]: I0124 00:07:47.530407 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4c6e9ec6352d1187d46c58912b6428fa13f7ab2ce0d3deb4c549df4b57bec338"} Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.539871 4676 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="1b050fc2d0d0d17898728982ae6d475c683cf41b177d47ea2a2be452323d9f0b" exitCode=0 Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.540017 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"1b050fc2d0d0d17898728982ae6d475c683cf41b177d47ea2a2be452323d9f0b"} Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.540312 4676 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.540348 4676 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.541683 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.542028 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:48 crc kubenswrapper[4676]: E0124 00:07:48.542097 4676 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.544905 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.544999 4676 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d" exitCode=1 Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.545052 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d"} Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.546157 4676 scope.go:117] "RemoveContainer" containerID="11b5a9331c7ed54da1e29daf0add6d4b15551929d37f1216b451e13b7d5ea94d" Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.546673 4676 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.547243 4676 status_manager.go:851] "Failed to get status for pod" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:48 crc kubenswrapper[4676]: I0124 00:07:48.547728 4676 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.27:6443: connect: connection refused" Jan 24 00:07:49 crc kubenswrapper[4676]: I0124 00:07:49.551832 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9af73b52d1d7410a18d6167074ecb96f5501caa21121bdc68a8da60db3dcb9cd"} Jan 24 00:07:49 crc kubenswrapper[4676]: I0124 00:07:49.552700 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"60b6dde5f3cba077c5ac3a95a9c47d9a96a1d09924ec7d461c9e18975ec769d0"} Jan 24 00:07:49 crc kubenswrapper[4676]: I0124 00:07:49.552721 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c5fa1473a44a3285dfd2dbd329fc5ad732f136d14f10cb3a0d3211302b2a13d3"} Jan 24 00:07:49 crc kubenswrapper[4676]: I0124 00:07:49.554739 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 24 00:07:49 crc kubenswrapper[4676]: I0124 00:07:49.554813 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d7b9fc46f06e374dfc7e4e9a0230fb3fa43bf3f2fbe5d4102849e87dc273f97c"} Jan 24 00:07:49 crc kubenswrapper[4676]: I0124 00:07:49.632399 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 24 00:07:49 crc kubenswrapper[4676]: I0124 00:07:49.637041 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 24 00:07:50 crc kubenswrapper[4676]: I0124 00:07:50.562367 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cd15469e4299972ae731dcd1ab90f9776eb75f14f17e8bdd477b3ea8146a25c6"} Jan 24 00:07:50 crc kubenswrapper[4676]: I0124 00:07:50.562477 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e02a952e3df13febc6f27d226d50c7693b0e26800feaaecd6de3534fa13f4e1f"} Jan 24 00:07:50 crc kubenswrapper[4676]: I0124 00:07:50.562511 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 24 00:07:50 crc kubenswrapper[4676]: I0124 00:07:50.562650 4676 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:50 crc kubenswrapper[4676]: I0124 00:07:50.562672 4676 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:52 crc kubenswrapper[4676]: I0124 00:07:52.280069 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:52 crc kubenswrapper[4676]: I0124 00:07:52.280411 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:52 crc kubenswrapper[4676]: I0124 00:07:52.284948 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:55 crc kubenswrapper[4676]: I0124 00:07:55.636159 4676 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:56 crc kubenswrapper[4676]: I0124 00:07:56.268291 4676 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ca258140-5ed5-4682-9321-00fc4c664028" Jan 24 00:07:56 crc kubenswrapper[4676]: I0124 00:07:56.608452 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:07:56 crc kubenswrapper[4676]: I0124 00:07:56.608623 4676 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:56 crc kubenswrapper[4676]: I0124 00:07:56.608653 4676 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:56 crc kubenswrapper[4676]: I0124 00:07:56.614100 4676 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ca258140-5ed5-4682-9321-00fc4c664028" Jan 24 00:07:57 crc kubenswrapper[4676]: I0124 00:07:57.614147 4676 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:57 crc kubenswrapper[4676]: I0124 00:07:57.614193 4676 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="653e6c74-9f8e-4c5f-b101-5b8da2e962ca" Jan 24 00:07:57 crc kubenswrapper[4676]: I0124 00:07:57.617327 4676 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ca258140-5ed5-4682-9321-00fc4c664028" Jan 24 00:08:05 crc kubenswrapper[4676]: I0124 00:08:05.480194 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 24 00:08:05 crc kubenswrapper[4676]: I0124 00:08:05.706498 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 24 00:08:05 crc kubenswrapper[4676]: I0124 00:08:05.750858 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 24 00:08:05 crc kubenswrapper[4676]: I0124 00:08:05.915058 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 24 00:08:05 crc kubenswrapper[4676]: I0124 00:08:05.976854 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 24 00:08:06 crc kubenswrapper[4676]: I0124 00:08:06.131354 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 24 00:08:06 crc kubenswrapper[4676]: I0124 00:08:06.591610 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 24 00:08:06 crc kubenswrapper[4676]: I0124 00:08:06.837559 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 24 00:08:06 crc kubenswrapper[4676]: I0124 00:08:06.895118 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 24 00:08:06 crc kubenswrapper[4676]: I0124 00:08:06.901593 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.024844 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.442732 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.468990 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.555352 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.603633 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.761491 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.806526 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.812512 4676 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.863714 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.874001 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.918710 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 24 00:08:07 crc kubenswrapper[4676]: I0124 00:08:07.995696 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.002500 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.092976 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.115127 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.128915 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.289845 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.400954 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.505456 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.701655 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.730236 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.816472 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.827234 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.862148 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.889535 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.920011 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 24 00:08:08 crc kubenswrapper[4676]: I0124 00:08:08.995552 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.158591 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.241832 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.326991 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.345937 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.384184 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.404452 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.426790 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.491974 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.540021 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.543009 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.651227 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.657855 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.780498 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.798019 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.817121 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 24 00:08:09 crc kubenswrapper[4676]: I0124 00:08:09.878222 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.001122 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.062425 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.087315 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.154592 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.211603 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.337362 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.392972 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.423032 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.452922 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.489665 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.641879 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.642516 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.678657 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.733231 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.733977 4676 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.763513 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.767918 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.814825 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.864366 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.893451 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 24 00:08:10 crc kubenswrapper[4676]: I0124 00:08:10.938084 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.114636 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.231561 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.233206 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.240681 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.283851 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.289566 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.365613 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.401155 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.415408 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.423253 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.437930 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.463848 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.517527 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.632469 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.819813 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.840273 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.887432 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.919887 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.953952 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 24 00:08:11 crc kubenswrapper[4676]: I0124 00:08:11.955648 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.015104 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.036759 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.068044 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.089414 4676 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.114402 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.131213 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.160346 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.166997 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.195006 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.253985 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.308007 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.378565 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.447670 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.449290 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.537225 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.603108 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.608140 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.658170 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.740873 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.829414 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.911894 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.923937 4676 patch_prober.go:28] interesting pod/console-operator-58897d9998-x649d container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.924002 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-x649d" podUID="32de8698-4bd5-4154-92a3-76930504a72d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.924471 4676 patch_prober.go:28] interesting pod/console-operator-58897d9998-x649d container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.924546 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-x649d" podUID="32de8698-4bd5-4154-92a3-76930504a72d" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:08:12 crc kubenswrapper[4676]: I0124 00:08:12.981284 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.029680 4676 patch_prober.go:28] interesting pod/console-f9d7485db-g2smk container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.029758 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-g2smk" podUID="5cce043a-2f1b-4f48-967e-c48a00cfe1a6" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.033820 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.086934 4676 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-n6vx5 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.087123 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" podUID="fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.086903 4676 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-n6vx5 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.087570 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" podUID="fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.113574 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.128634 4676 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-ft4kq container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": context deadline exceeded" start-of-body= Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.128739 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-ft4kq" podUID="e6351c23-e315-4c92-a467-380da403d3c4" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": context deadline exceeded" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.179961 4676 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.182748 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=40.182727886 podStartE2EDuration="40.182727886s" podCreationTimestamp="2026-01-24 00:07:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:07:55.670434198 +0000 UTC m=+259.700405229" watchObservedRunningTime="2026-01-24 00:08:13.182727886 +0000 UTC m=+277.212698887" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.185643 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.185700 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.194321 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.194900 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.208538 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.233488 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.234599 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.242215 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.337741 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.340214 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.340189025 podStartE2EDuration="18.340189025s" podCreationTimestamp="2026-01-24 00:07:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:08:13.217423137 +0000 UTC m=+277.247394138" watchObservedRunningTime="2026-01-24 00:08:13.340189025 +0000 UTC m=+277.370160026" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.418116 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.445598 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.561236 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.664903 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.730102 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.745617 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.763913 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.881895 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.956317 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 24 00:08:13 crc kubenswrapper[4676]: I0124 00:08:13.978271 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.223529 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.396971 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.543309 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.606128 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.615182 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.659423 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.696066 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.702371 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.771101 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 24 00:08:14 crc kubenswrapper[4676]: I0124 00:08:14.871113 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.056808 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.085440 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.151627 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.173998 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.280625 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.332215 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.360555 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.365230 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.399911 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.431502 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.442023 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.500341 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.515648 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.713034 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.718731 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.724720 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.849564 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 24 00:08:15 crc kubenswrapper[4676]: I0124 00:08:15.859038 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.022224 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.031474 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.193752 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.196757 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.258815 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.448582 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.545652 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.588203 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.626903 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.658888 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.677587 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.746415 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.807205 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.824206 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.890334 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 24 00:08:16 crc kubenswrapper[4676]: I0124 00:08:16.993601 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.082513 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.106112 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.107014 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.139152 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.204764 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.234214 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.249358 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.269403 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.272418 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.280011 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.306626 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.308040 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.326097 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.364205 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.386571 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.520932 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.741788 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.777114 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.793502 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.835005 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 24 00:08:17 crc kubenswrapper[4676]: I0124 00:08:17.916402 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:17.947995 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.050843 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.082542 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.108590 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.146603 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.183966 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.185454 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.262239 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.295521 4676 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.295861 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.295880 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://31446979da9b2814d80beb1cd72c15277eaaac69a2a2c9676d049a6a8acf10c5" gracePeriod=5 Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.491147 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.568722 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.651607 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.682943 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.708611 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.851103 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.914638 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.974243 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 24 00:08:18 crc kubenswrapper[4676]: I0124 00:08:18.981936 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.012699 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.092973 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.124096 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.163885 4676 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.197165 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.289569 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.293610 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.398417 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.467555 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.472722 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.493369 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.590874 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.608294 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.666355 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.676583 4676 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.680920 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.728448 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.729068 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.747541 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.941860 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.998007 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 24 00:08:19 crc kubenswrapper[4676]: I0124 00:08:19.998025 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 24 00:08:20 crc kubenswrapper[4676]: I0124 00:08:20.077928 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 24 00:08:20 crc kubenswrapper[4676]: I0124 00:08:20.131679 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 24 00:08:20 crc kubenswrapper[4676]: I0124 00:08:20.168624 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 24 00:08:20 crc kubenswrapper[4676]: I0124 00:08:20.368915 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 24 00:08:20 crc kubenswrapper[4676]: I0124 00:08:20.441033 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 24 00:08:20 crc kubenswrapper[4676]: I0124 00:08:20.597066 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 24 00:08:20 crc kubenswrapper[4676]: I0124 00:08:20.934503 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 24 00:08:21 crc kubenswrapper[4676]: I0124 00:08:21.289050 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 24 00:08:21 crc kubenswrapper[4676]: I0124 00:08:21.735730 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.240667 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.365472 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.497367 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.629394 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lxs47"] Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.629674 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lxs47" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerName="registry-server" containerID="cri-o://c53d6e9162fb162b724098a40f36dec02c2bb75d99128d9311568c19c9c217a4" gracePeriod=30 Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.634647 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pnqjd"] Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.634999 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pnqjd" podUID="891b78f7-509c-4e8d-b846-52881396a64d" containerName="registry-server" containerID="cri-o://fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461" gracePeriod=30 Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.645924 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rxmrn"] Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.646255 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rxmrn" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerName="registry-server" containerID="cri-o://eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0" gracePeriod=30 Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.654064 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rsw66"] Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.654264 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" podUID="7af03157-ee92-4e72-a775-acaeabb73e65" containerName="marketplace-operator" containerID="cri-o://a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c" gracePeriod=30 Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.665945 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7nsm"] Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.666237 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l7nsm" podUID="920c325a-f36b-4162-9d37-ea88124be938" containerName="registry-server" containerID="cri-o://cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211" gracePeriod=30 Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.689331 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-645jr"] Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.689757 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-645jr" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="registry-server" containerID="cri-o://1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449" gracePeriod=30 Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.719742 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kwptt"] Jan 24 00:08:22 crc kubenswrapper[4676]: E0124 00:08:22.719956 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" containerName="installer" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.719971 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" containerName="installer" Jan 24 00:08:22 crc kubenswrapper[4676]: E0124 00:08:22.719986 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.719995 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.720095 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="c280bb7c-e926-46c9-bb2c-d88c64634673" containerName="installer" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.720110 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.720541 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.732409 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kwptt"] Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.740694 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 24 00:08:22 crc kubenswrapper[4676]: E0124 00:08:22.778428 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449 is running failed: container process not found" containerID="1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449" cmd=["grpc_health_probe","-addr=:50051"] Jan 24 00:08:22 crc kubenswrapper[4676]: E0124 00:08:22.779460 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449 is running failed: container process not found" containerID="1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449" cmd=["grpc_health_probe","-addr=:50051"] Jan 24 00:08:22 crc kubenswrapper[4676]: E0124 00:08:22.786794 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449 is running failed: container process not found" containerID="1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449" cmd=["grpc_health_probe","-addr=:50051"] Jan 24 00:08:22 crc kubenswrapper[4676]: E0124 00:08:22.786839 4676 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-645jr" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="registry-server" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.794691 4676 generic.go:334] "Generic (PLEG): container finished" podID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerID="c53d6e9162fb162b724098a40f36dec02c2bb75d99128d9311568c19c9c217a4" exitCode=0 Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.794746 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxs47" event={"ID":"592859d8-1f7e-4e35-acc9-635e130ad2d2","Type":"ContainerDied","Data":"c53d6e9162fb162b724098a40f36dec02c2bb75d99128d9311568c19c9c217a4"} Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.851820 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pnqjd"] Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.922075 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzqgz\" (UniqueName: \"kubernetes.io/projected/8db31db6-2c7c-4688-89c7-328024cd7003-kube-api-access-zzqgz\") pod \"marketplace-operator-79b997595-kwptt\" (UID: \"8db31db6-2c7c-4688-89c7-328024cd7003\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.922170 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8db31db6-2c7c-4688-89c7-328024cd7003-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kwptt\" (UID: \"8db31db6-2c7c-4688-89c7-328024cd7003\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:22 crc kubenswrapper[4676]: I0124 00:08:22.922260 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8db31db6-2c7c-4688-89c7-328024cd7003-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kwptt\" (UID: \"8db31db6-2c7c-4688-89c7-328024cd7003\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.023191 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzqgz\" (UniqueName: \"kubernetes.io/projected/8db31db6-2c7c-4688-89c7-328024cd7003-kube-api-access-zzqgz\") pod \"marketplace-operator-79b997595-kwptt\" (UID: \"8db31db6-2c7c-4688-89c7-328024cd7003\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.023239 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8db31db6-2c7c-4688-89c7-328024cd7003-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kwptt\" (UID: \"8db31db6-2c7c-4688-89c7-328024cd7003\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.023428 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8db31db6-2c7c-4688-89c7-328024cd7003-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kwptt\" (UID: \"8db31db6-2c7c-4688-89c7-328024cd7003\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.024682 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8db31db6-2c7c-4688-89c7-328024cd7003-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-kwptt\" (UID: \"8db31db6-2c7c-4688-89c7-328024cd7003\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.032960 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8db31db6-2c7c-4688-89c7-328024cd7003-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-kwptt\" (UID: \"8db31db6-2c7c-4688-89c7-328024cd7003\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.038707 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzqgz\" (UniqueName: \"kubernetes.io/projected/8db31db6-2c7c-4688-89c7-328024cd7003-kube-api-access-zzqgz\") pod \"marketplace-operator-79b997595-kwptt\" (UID: \"8db31db6-2c7c-4688-89c7-328024cd7003\") " pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.089554 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.179473 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.181173 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.185334 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.189555 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.194741 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.225187 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzqlz\" (UniqueName: \"kubernetes.io/projected/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-kube-api-access-hzqlz\") pod \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.226295 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-catalog-content\") pod \"592859d8-1f7e-4e35-acc9-635e130ad2d2\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.226425 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-catalog-content\") pod \"920c325a-f36b-4162-9d37-ea88124be938\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.226536 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g5kz\" (UniqueName: \"kubernetes.io/projected/10798203-b391-4a87-98a7-b41db2bbb0e2-kube-api-access-8g5kz\") pod \"10798203-b391-4a87-98a7-b41db2bbb0e2\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.227907 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-catalog-content\") pod \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.228066 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49sw6\" (UniqueName: \"kubernetes.io/projected/891b78f7-509c-4e8d-b846-52881396a64d-kube-api-access-49sw6\") pod \"891b78f7-509c-4e8d-b846-52881396a64d\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.228211 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4p5l\" (UniqueName: \"kubernetes.io/projected/920c325a-f36b-4162-9d37-ea88124be938-kube-api-access-n4p5l\") pod \"920c325a-f36b-4162-9d37-ea88124be938\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.228417 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-utilities\") pod \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\" (UID: \"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.228585 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjhrt\" (UniqueName: \"kubernetes.io/projected/592859d8-1f7e-4e35-acc9-635e130ad2d2-kube-api-access-tjhrt\") pod \"592859d8-1f7e-4e35-acc9-635e130ad2d2\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.228754 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-operator-metrics\") pod \"7af03157-ee92-4e72-a775-acaeabb73e65\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.229587 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-catalog-content\") pod \"891b78f7-509c-4e8d-b846-52881396a64d\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.229722 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-catalog-content\") pod \"10798203-b391-4a87-98a7-b41db2bbb0e2\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.229829 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-utilities\") pod \"920c325a-f36b-4162-9d37-ea88124be938\" (UID: \"920c325a-f36b-4162-9d37-ea88124be938\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.231334 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-utilities" (OuterVolumeSpecName: "utilities") pod "920c325a-f36b-4162-9d37-ea88124be938" (UID: "920c325a-f36b-4162-9d37-ea88124be938"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.234332 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-utilities" (OuterVolumeSpecName: "utilities") pod "0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" (UID: "0a4d7c63-cff0-4408-9cb6-450f3ebc53dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.239652 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10798203-b391-4a87-98a7-b41db2bbb0e2-kube-api-access-8g5kz" (OuterVolumeSpecName: "kube-api-access-8g5kz") pod "10798203-b391-4a87-98a7-b41db2bbb0e2" (UID: "10798203-b391-4a87-98a7-b41db2bbb0e2"). InnerVolumeSpecName "kube-api-access-8g5kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.245324 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/891b78f7-509c-4e8d-b846-52881396a64d-kube-api-access-49sw6" (OuterVolumeSpecName: "kube-api-access-49sw6") pod "891b78f7-509c-4e8d-b846-52881396a64d" (UID: "891b78f7-509c-4e8d-b846-52881396a64d"). InnerVolumeSpecName "kube-api-access-49sw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.245572 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/920c325a-f36b-4162-9d37-ea88124be938-kube-api-access-n4p5l" (OuterVolumeSpecName: "kube-api-access-n4p5l") pod "920c325a-f36b-4162-9d37-ea88124be938" (UID: "920c325a-f36b-4162-9d37-ea88124be938"). InnerVolumeSpecName "kube-api-access-n4p5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.245868 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "7af03157-ee92-4e72-a775-acaeabb73e65" (UID: "7af03157-ee92-4e72-a775-acaeabb73e65"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.250048 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/592859d8-1f7e-4e35-acc9-635e130ad2d2-kube-api-access-tjhrt" (OuterVolumeSpecName: "kube-api-access-tjhrt") pod "592859d8-1f7e-4e35-acc9-635e130ad2d2" (UID: "592859d8-1f7e-4e35-acc9-635e130ad2d2"). InnerVolumeSpecName "kube-api-access-tjhrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.256130 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-kube-api-access-hzqlz" (OuterVolumeSpecName: "kube-api-access-hzqlz") pod "0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" (UID: "0a4d7c63-cff0-4408-9cb6-450f3ebc53dd"). InnerVolumeSpecName "kube-api-access-hzqlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.279105 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "920c325a-f36b-4162-9d37-ea88124be938" (UID: "920c325a-f36b-4162-9d37-ea88124be938"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.290658 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.296619 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" (UID: "0a4d7c63-cff0-4408-9cb6-450f3ebc53dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.312271 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "592859d8-1f7e-4e35-acc9-635e130ad2d2" (UID: "592859d8-1f7e-4e35-acc9-635e130ad2d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.316279 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "891b78f7-509c-4e8d-b846-52881396a64d" (UID: "891b78f7-509c-4e8d-b846-52881396a64d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.330881 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-utilities\") pod \"592859d8-1f7e-4e35-acc9-635e130ad2d2\" (UID: \"592859d8-1f7e-4e35-acc9-635e130ad2d2\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.330932 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm4tl\" (UniqueName: \"kubernetes.io/projected/7af03157-ee92-4e72-a775-acaeabb73e65-kube-api-access-jm4tl\") pod \"7af03157-ee92-4e72-a775-acaeabb73e65\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.330953 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-utilities\") pod \"10798203-b391-4a87-98a7-b41db2bbb0e2\" (UID: \"10798203-b391-4a87-98a7-b41db2bbb0e2\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.330973 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-utilities\") pod \"891b78f7-509c-4e8d-b846-52881396a64d\" (UID: \"891b78f7-509c-4e8d-b846-52881396a64d\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331017 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-trusted-ca\") pod \"7af03157-ee92-4e72-a775-acaeabb73e65\" (UID: \"7af03157-ee92-4e72-a775-acaeabb73e65\") " Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331295 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzqlz\" (UniqueName: \"kubernetes.io/projected/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-kube-api-access-hzqlz\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331306 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331315 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331324 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g5kz\" (UniqueName: \"kubernetes.io/projected/10798203-b391-4a87-98a7-b41db2bbb0e2-kube-api-access-8g5kz\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331332 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331710 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49sw6\" (UniqueName: \"kubernetes.io/projected/891b78f7-509c-4e8d-b846-52881396a64d-kube-api-access-49sw6\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331721 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4p5l\" (UniqueName: \"kubernetes.io/projected/920c325a-f36b-4162-9d37-ea88124be938-kube-api-access-n4p5l\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331730 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331739 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjhrt\" (UniqueName: \"kubernetes.io/projected/592859d8-1f7e-4e35-acc9-635e130ad2d2-kube-api-access-tjhrt\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331747 4676 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331756 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.331765 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/920c325a-f36b-4162-9d37-ea88124be938-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.333034 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-utilities" (OuterVolumeSpecName: "utilities") pod "592859d8-1f7e-4e35-acc9-635e130ad2d2" (UID: "592859d8-1f7e-4e35-acc9-635e130ad2d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.341145 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "7af03157-ee92-4e72-a775-acaeabb73e65" (UID: "7af03157-ee92-4e72-a775-acaeabb73e65"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.341350 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-utilities" (OuterVolumeSpecName: "utilities") pod "10798203-b391-4a87-98a7-b41db2bbb0e2" (UID: "10798203-b391-4a87-98a7-b41db2bbb0e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.341859 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-utilities" (OuterVolumeSpecName: "utilities") pod "891b78f7-509c-4e8d-b846-52881396a64d" (UID: "891b78f7-509c-4e8d-b846-52881396a64d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.344563 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7af03157-ee92-4e72-a775-acaeabb73e65-kube-api-access-jm4tl" (OuterVolumeSpecName: "kube-api-access-jm4tl") pod "7af03157-ee92-4e72-a775-acaeabb73e65" (UID: "7af03157-ee92-4e72-a775-acaeabb73e65"). InnerVolumeSpecName "kube-api-access-jm4tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.363470 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.412341 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "10798203-b391-4a87-98a7-b41db2bbb0e2" (UID: "10798203-b391-4a87-98a7-b41db2bbb0e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.432183 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.432206 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/592859d8-1f7e-4e35-acc9-635e130ad2d2-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.432217 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm4tl\" (UniqueName: \"kubernetes.io/projected/7af03157-ee92-4e72-a775-acaeabb73e65-kube-api-access-jm4tl\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.432227 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/10798203-b391-4a87-98a7-b41db2bbb0e2-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.432236 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/891b78f7-509c-4e8d-b846-52881396a64d-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.432245 4676 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7af03157-ee92-4e72-a775-acaeabb73e65-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.491161 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-kwptt"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.803181 4676 generic.go:334] "Generic (PLEG): container finished" podID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerID="eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0" exitCode=0 Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.803237 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rxmrn" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.803287 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxmrn" event={"ID":"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd","Type":"ContainerDied","Data":"eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.803348 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rxmrn" event={"ID":"0a4d7c63-cff0-4408-9cb6-450f3ebc53dd","Type":"ContainerDied","Data":"e7b1f0797950f02a9f5633d0b1bb88fc987a1236bb934934be7f6364e391819f"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.803387 4676 scope.go:117] "RemoveContainer" containerID="eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.806332 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" event={"ID":"8db31db6-2c7c-4688-89c7-328024cd7003","Type":"ContainerStarted","Data":"f747fa388ee96adbcd130ec185be58f0f23b427ed1052f15f08077aca9cb4bec"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.806362 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" event={"ID":"8db31db6-2c7c-4688-89c7-328024cd7003","Type":"ContainerStarted","Data":"4ab3c7b1ae83baac88eadf6edce3d31ceb98ea3b39880093b168a5c27926ed80"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.806566 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.809130 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lxs47" event={"ID":"592859d8-1f7e-4e35-acc9-635e130ad2d2","Type":"ContainerDied","Data":"13ab7acc7e8b9f7cc9283df9a7fe0c7c12a31d5e17ec678008c5409224450c7c"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.809248 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lxs47" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.813474 4676 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-kwptt container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" start-of-body= Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.813527 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" podUID="8db31db6-2c7c-4688-89c7-328024cd7003" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.58:8080/healthz\": dial tcp 10.217.0.58:8080: connect: connection refused" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.819595 4676 generic.go:334] "Generic (PLEG): container finished" podID="920c325a-f36b-4162-9d37-ea88124be938" containerID="cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211" exitCode=0 Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.819651 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l7nsm" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.819669 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7nsm" event={"ID":"920c325a-f36b-4162-9d37-ea88124be938","Type":"ContainerDied","Data":"cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.819702 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l7nsm" event={"ID":"920c325a-f36b-4162-9d37-ea88124be938","Type":"ContainerDied","Data":"c30184174638423d11a5d0c24917e087282eea0233650b891df20e30923de797"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.822511 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.822551 4676 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="31446979da9b2814d80beb1cd72c15277eaaac69a2a2c9676d049a6a8acf10c5" exitCode=137 Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.829251 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" podStartSLOduration=1.829235267 podStartE2EDuration="1.829235267s" podCreationTimestamp="2026-01-24 00:08:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:08:23.827639375 +0000 UTC m=+287.857610376" watchObservedRunningTime="2026-01-24 00:08:23.829235267 +0000 UTC m=+287.859206268" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.837702 4676 generic.go:334] "Generic (PLEG): container finished" podID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerID="1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449" exitCode=0 Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.837826 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645jr" event={"ID":"10798203-b391-4a87-98a7-b41db2bbb0e2","Type":"ContainerDied","Data":"1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.837874 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-645jr" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.837899 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645jr" event={"ID":"10798203-b391-4a87-98a7-b41db2bbb0e2","Type":"ContainerDied","Data":"bc7d200b5a89367114e48ce959705259b571653e291943569fe69329c0b533b4"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.844468 4676 generic.go:334] "Generic (PLEG): container finished" podID="7af03157-ee92-4e72-a775-acaeabb73e65" containerID="a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c" exitCode=0 Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.844565 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" event={"ID":"7af03157-ee92-4e72-a775-acaeabb73e65","Type":"ContainerDied","Data":"a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.844591 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" event={"ID":"7af03157-ee92-4e72-a775-acaeabb73e65","Type":"ContainerDied","Data":"3e3f107eb805996a69d5ec1a562a4e16c0b2393227f3f9858f21b04b163f17b3"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.844675 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rsw66" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.852661 4676 scope.go:117] "RemoveContainer" containerID="932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.877616 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rxmrn"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.887790 4676 generic.go:334] "Generic (PLEG): container finished" podID="891b78f7-509c-4e8d-b846-52881396a64d" containerID="fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461" exitCode=0 Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.887954 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnqjd" event={"ID":"891b78f7-509c-4e8d-b846-52881396a64d","Type":"ContainerDied","Data":"fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.888063 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pnqjd" event={"ID":"891b78f7-509c-4e8d-b846-52881396a64d","Type":"ContainerDied","Data":"fe1008d35989eb8d8f31d34ea59eda3fc8a16f8472a8718f92a963e88a25a092"} Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.888420 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pnqjd" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.888823 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rxmrn"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.898851 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lxs47"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.903446 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lxs47"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.909811 4676 scope.go:117] "RemoveContainer" containerID="da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.910246 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7nsm"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.914584 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l7nsm"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.921394 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.921452 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.929253 4676 scope.go:117] "RemoveContainer" containerID="eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0" Jan 24 00:08:23 crc kubenswrapper[4676]: E0124 00:08:23.930080 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0\": container with ID starting with eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0 not found: ID does not exist" containerID="eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.930122 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0"} err="failed to get container status \"eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0\": rpc error: code = NotFound desc = could not find container \"eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0\": container with ID starting with eaf3e81af38d0f1a8b5ae4e7e29a2416e79e9d3282c5a5d4ac1c2360695efab0 not found: ID does not exist" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.930148 4676 scope.go:117] "RemoveContainer" containerID="932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c" Jan 24 00:08:23 crc kubenswrapper[4676]: E0124 00:08:23.930450 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c\": container with ID starting with 932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c not found: ID does not exist" containerID="932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.930471 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c"} err="failed to get container status \"932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c\": rpc error: code = NotFound desc = could not find container \"932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c\": container with ID starting with 932c49b1763ff02b2ff12475c95193c10bded4a7c1630e37d48894f5fd7a938c not found: ID does not exist" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.930485 4676 scope.go:117] "RemoveContainer" containerID="da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060" Jan 24 00:08:23 crc kubenswrapper[4676]: E0124 00:08:23.930773 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060\": container with ID starting with da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060 not found: ID does not exist" containerID="da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.930796 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060"} err="failed to get container status \"da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060\": rpc error: code = NotFound desc = could not find container \"da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060\": container with ID starting with da65e02cce8de032757324d83329948b4cdaba38f23cea059f4dfacc20dbe060 not found: ID does not exist" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.930919 4676 scope.go:117] "RemoveContainer" containerID="c53d6e9162fb162b724098a40f36dec02c2bb75d99128d9311568c19c9c217a4" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.931788 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-645jr"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.939534 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-645jr"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.947930 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pnqjd"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.952430 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pnqjd"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.954742 4676 scope.go:117] "RemoveContainer" containerID="d84cf120443faf814099d3b57961bbb33e517c1b14f12a7cc4e82668a41edcc9" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.958299 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rsw66"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.961550 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rsw66"] Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.982619 4676 scope.go:117] "RemoveContainer" containerID="5f55960a6c743513af98906e24b7fac1e853e9cbe06b95180d6ac6353a01cca4" Jan 24 00:08:23 crc kubenswrapper[4676]: I0124 00:08:23.995866 4676 scope.go:117] "RemoveContainer" containerID="cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.010415 4676 scope.go:117] "RemoveContainer" containerID="c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.026151 4676 scope.go:117] "RemoveContainer" containerID="1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.037657 4676 scope.go:117] "RemoveContainer" containerID="cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.038402 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211\": container with ID starting with cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211 not found: ID does not exist" containerID="cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.038444 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211"} err="failed to get container status \"cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211\": rpc error: code = NotFound desc = could not find container \"cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211\": container with ID starting with cec35737af48fd560021306cb0c7587b6f69b93f483a188d615bbf073ab3c211 not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.038479 4676 scope.go:117] "RemoveContainer" containerID="c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.038894 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.038943 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.038970 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039016 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039050 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039102 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039271 4676 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039050 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039067 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039317 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.039386 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b\": container with ID starting with c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b not found: ID does not exist" containerID="c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039407 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b"} err="failed to get container status \"c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b\": rpc error: code = NotFound desc = could not find container \"c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b\": container with ID starting with c367a0daec440166091e52dccf8eb465443af958a90797e10fac9e024d33ce5b not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039427 4676 scope.go:117] "RemoveContainer" containerID="1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.039722 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3\": container with ID starting with 1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3 not found: ID does not exist" containerID="1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039745 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3"} err="failed to get container status \"1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3\": rpc error: code = NotFound desc = could not find container \"1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3\": container with ID starting with 1bdebd4e7e3f1f39e0a22ca3c50b2e8f9a237356bdbbca8ae01f28c88f7a72d3 not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.039758 4676 scope.go:117] "RemoveContainer" containerID="1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.044675 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.051193 4676 scope.go:117] "RemoveContainer" containerID="664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.063291 4676 scope.go:117] "RemoveContainer" containerID="eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.076162 4676 scope.go:117] "RemoveContainer" containerID="1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.076707 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449\": container with ID starting with 1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449 not found: ID does not exist" containerID="1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.076869 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449"} err="failed to get container status \"1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449\": rpc error: code = NotFound desc = could not find container \"1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449\": container with ID starting with 1884c81b387764c68fe7f812469ffb55cdc424867d5432e35c47580ec4c93449 not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.077041 4676 scope.go:117] "RemoveContainer" containerID="664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.077593 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233\": container with ID starting with 664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233 not found: ID does not exist" containerID="664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.077619 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233"} err="failed to get container status \"664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233\": rpc error: code = NotFound desc = could not find container \"664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233\": container with ID starting with 664b5bc200aa1b61de8d525097a5da8a9ecda54e783284807e9107a8e1a20233 not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.077634 4676 scope.go:117] "RemoveContainer" containerID="eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.077986 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0\": container with ID starting with eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0 not found: ID does not exist" containerID="eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.078128 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0"} err="failed to get container status \"eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0\": rpc error: code = NotFound desc = could not find container \"eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0\": container with ID starting with eb96ef854f131482318ba1e9541dfd7cdefbb362e6e310e27326cdd1db4c3ea0 not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.078257 4676 scope.go:117] "RemoveContainer" containerID="a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.093809 4676 scope.go:117] "RemoveContainer" containerID="a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.094228 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c\": container with ID starting with a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c not found: ID does not exist" containerID="a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.094256 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c"} err="failed to get container status \"a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c\": rpc error: code = NotFound desc = could not find container \"a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c\": container with ID starting with a20606a37ebc8d77d9e63ac6ed5eaa05b5b45e9454c3cc98154b304c79f1827c not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.094282 4676 scope.go:117] "RemoveContainer" containerID="fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.106696 4676 scope.go:117] "RemoveContainer" containerID="ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.118228 4676 scope.go:117] "RemoveContainer" containerID="5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.132115 4676 scope.go:117] "RemoveContainer" containerID="fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.132552 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461\": container with ID starting with fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461 not found: ID does not exist" containerID="fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.132650 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461"} err="failed to get container status \"fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461\": rpc error: code = NotFound desc = could not find container \"fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461\": container with ID starting with fba973515d63eaeb97481aaf2d27398145972a0521a5d48402281dfaca486461 not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.132731 4676 scope.go:117] "RemoveContainer" containerID="ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.133025 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903\": container with ID starting with ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903 not found: ID does not exist" containerID="ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.133050 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903"} err="failed to get container status \"ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903\": rpc error: code = NotFound desc = could not find container \"ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903\": container with ID starting with ec0016b4a3649cb82109f7041ce0d5dd5501d9ceb2f93c0c5e1205592a82d903 not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.133069 4676 scope.go:117] "RemoveContainer" containerID="5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0" Jan 24 00:08:24 crc kubenswrapper[4676]: E0124 00:08:24.133315 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0\": container with ID starting with 5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0 not found: ID does not exist" containerID="5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.133334 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0"} err="failed to get container status \"5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0\": rpc error: code = NotFound desc = could not find container \"5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0\": container with ID starting with 5f5e9794469ec06bd85502c50865410e6c1b4c331eb790f8073e4c59c7182bb0 not found: ID does not exist" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.140104 4676 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.140128 4676 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.140139 4676 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.140148 4676 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.261878 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" path="/var/lib/kubelet/pods/0a4d7c63-cff0-4408-9cb6-450f3ebc53dd/volumes" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.262669 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" path="/var/lib/kubelet/pods/10798203-b391-4a87-98a7-b41db2bbb0e2/volumes" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.263450 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" path="/var/lib/kubelet/pods/592859d8-1f7e-4e35-acc9-635e130ad2d2/volumes" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.264846 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7af03157-ee92-4e72-a775-acaeabb73e65" path="/var/lib/kubelet/pods/7af03157-ee92-4e72-a775-acaeabb73e65/volumes" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.265540 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="891b78f7-509c-4e8d-b846-52881396a64d" path="/var/lib/kubelet/pods/891b78f7-509c-4e8d-b846-52881396a64d/volumes" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.267118 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="920c325a-f36b-4162-9d37-ea88124be938" path="/var/lib/kubelet/pods/920c325a-f36b-4162-9d37-ea88124be938/volumes" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.267790 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.268090 4676 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.279027 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.279250 4676 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="0c058973-4f5f-43eb-a86b-a70d5b899ad6" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.281786 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.281804 4676 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="0c058973-4f5f-43eb-a86b-a70d5b899ad6" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.906477 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.906590 4676 scope.go:117] "RemoveContainer" containerID="31446979da9b2814d80beb1cd72c15277eaaac69a2a2c9676d049a6a8acf10c5" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.906621 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 24 00:08:24 crc kubenswrapper[4676]: I0124 00:08:24.915870 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-kwptt" Jan 24 00:08:36 crc kubenswrapper[4676]: I0124 00:08:36.124283 4676 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 24 00:08:47 crc kubenswrapper[4676]: I0124 00:08:47.355490 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qzmmz"] Jan 24 00:08:47 crc kubenswrapper[4676]: I0124 00:08:47.357925 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" podUID="b37a2847-a94a-4a0c-b092-1ed7155a2d35" containerName="controller-manager" containerID="cri-o://75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed" gracePeriod=30 Jan 24 00:08:47 crc kubenswrapper[4676]: I0124 00:08:47.454244 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5"] Jan 24 00:08:47 crc kubenswrapper[4676]: I0124 00:08:47.454903 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" podUID="fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" containerName="route-controller-manager" containerID="cri-o://f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa" gracePeriod=30 Jan 24 00:08:48 crc kubenswrapper[4676]: I0124 00:08:48.908385 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.011618 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.044887 4676 generic.go:334] "Generic (PLEG): container finished" podID="fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" containerID="f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa" exitCode=0 Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.044941 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" event={"ID":"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93","Type":"ContainerDied","Data":"f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa"} Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.045001 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" event={"ID":"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93","Type":"ContainerDied","Data":"caafeb67c7aefa20810226a8341333cb55d2bc2743763b95b6989d1d99cdf707"} Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.045027 4676 scope.go:117] "RemoveContainer" containerID="f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.045030 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.046758 4676 generic.go:334] "Generic (PLEG): container finished" podID="b37a2847-a94a-4a0c-b092-1ed7155a2d35" containerID="75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed" exitCode=0 Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.046801 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" event={"ID":"b37a2847-a94a-4a0c-b092-1ed7155a2d35","Type":"ContainerDied","Data":"75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed"} Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.046818 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" event={"ID":"b37a2847-a94a-4a0c-b092-1ed7155a2d35","Type":"ContainerDied","Data":"63dfdd35be2732ab824c6cbb0405da330ff5fefbf093c8f9f68db2d40a451da2"} Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.046833 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qzmmz" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.047583 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-config\") pod \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.047625 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-proxy-ca-bundles\") pod \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.047683 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv9mz\" (UniqueName: \"kubernetes.io/projected/b37a2847-a94a-4a0c-b092-1ed7155a2d35-kube-api-access-nv9mz\") pod \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.047711 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37a2847-a94a-4a0c-b092-1ed7155a2d35-serving-cert\") pod \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.047766 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-client-ca\") pod \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\" (UID: \"b37a2847-a94a-4a0c-b092-1ed7155a2d35\") " Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.048619 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-client-ca" (OuterVolumeSpecName: "client-ca") pod "b37a2847-a94a-4a0c-b092-1ed7155a2d35" (UID: "b37a2847-a94a-4a0c-b092-1ed7155a2d35"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.051802 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b37a2847-a94a-4a0c-b092-1ed7155a2d35" (UID: "b37a2847-a94a-4a0c-b092-1ed7155a2d35"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.054036 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-config" (OuterVolumeSpecName: "config") pod "b37a2847-a94a-4a0c-b092-1ed7155a2d35" (UID: "b37a2847-a94a-4a0c-b092-1ed7155a2d35"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.054103 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b37a2847-a94a-4a0c-b092-1ed7155a2d35-kube-api-access-nv9mz" (OuterVolumeSpecName: "kube-api-access-nv9mz") pod "b37a2847-a94a-4a0c-b092-1ed7155a2d35" (UID: "b37a2847-a94a-4a0c-b092-1ed7155a2d35"). InnerVolumeSpecName "kube-api-access-nv9mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.054734 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b37a2847-a94a-4a0c-b092-1ed7155a2d35-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b37a2847-a94a-4a0c-b092-1ed7155a2d35" (UID: "b37a2847-a94a-4a0c-b092-1ed7155a2d35"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.064447 4676 scope.go:117] "RemoveContainer" containerID="f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa" Jan 24 00:08:49 crc kubenswrapper[4676]: E0124 00:08:49.064841 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa\": container with ID starting with f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa not found: ID does not exist" containerID="f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.064878 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa"} err="failed to get container status \"f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa\": rpc error: code = NotFound desc = could not find container \"f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa\": container with ID starting with f4280b5e7bc19b43a820be872f8d3506b0ceb5bfbe0f82d0caa7fb857177ccaa not found: ID does not exist" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.064908 4676 scope.go:117] "RemoveContainer" containerID="75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.085710 4676 scope.go:117] "RemoveContainer" containerID="75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed" Jan 24 00:08:49 crc kubenswrapper[4676]: E0124 00:08:49.086331 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed\": container with ID starting with 75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed not found: ID does not exist" containerID="75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.086366 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed"} err="failed to get container status \"75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed\": rpc error: code = NotFound desc = could not find container \"75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed\": container with ID starting with 75da5a7bac808efea95403cb0f28f0e32a88fac01cca8335b0481d14b39d80ed not found: ID does not exist" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.148497 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-client-ca\") pod \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.148568 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-config\") pod \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.148596 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4cw9\" (UniqueName: \"kubernetes.io/projected/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-kube-api-access-h4cw9\") pod \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.148626 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-serving-cert\") pod \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\" (UID: \"fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93\") " Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.148828 4676 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-client-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.148840 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.148850 4676 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b37a2847-a94a-4a0c-b092-1ed7155a2d35-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.148861 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv9mz\" (UniqueName: \"kubernetes.io/projected/b37a2847-a94a-4a0c-b092-1ed7155a2d35-kube-api-access-nv9mz\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.148869 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b37a2847-a94a-4a0c-b092-1ed7155a2d35-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.149597 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-config" (OuterVolumeSpecName: "config") pod "fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" (UID: "fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.149594 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-client-ca" (OuterVolumeSpecName: "client-ca") pod "fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" (UID: "fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.152678 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" (UID: "fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.153667 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-kube-api-access-h4cw9" (OuterVolumeSpecName: "kube-api-access-h4cw9") pod "fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" (UID: "fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93"). InnerVolumeSpecName "kube-api-access-h4cw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.250326 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.250357 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4cw9\" (UniqueName: \"kubernetes.io/projected/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-kube-api-access-h4cw9\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.250398 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.250407 4676 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93-client-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.393793 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qzmmz"] Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.402677 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qzmmz"] Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.405987 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5"] Jan 24 00:08:49 crc kubenswrapper[4676]: I0124 00:08:49.408739 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-n6vx5"] Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.266435 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b37a2847-a94a-4a0c-b092-1ed7155a2d35" path="/var/lib/kubelet/pods/b37a2847-a94a-4a0c-b092-1ed7155a2d35/volumes" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.268087 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" path="/var/lib/kubelet/pods/fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93/volumes" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.373615 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9qdxp"] Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.373836 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="891b78f7-509c-4e8d-b846-52881396a64d" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.373850 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="891b78f7-509c-4e8d-b846-52881396a64d" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.373866 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.373875 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.373887 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.373896 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.373906 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="891b78f7-509c-4e8d-b846-52881396a64d" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.373915 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="891b78f7-509c-4e8d-b846-52881396a64d" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.373924 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.373931 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.373943 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920c325a-f36b-4162-9d37-ea88124be938" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.373952 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="920c325a-f36b-4162-9d37-ea88124be938" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.373962 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.373971 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.373983 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" containerName="route-controller-manager" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.373993 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" containerName="route-controller-manager" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374007 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374014 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374026 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920c325a-f36b-4162-9d37-ea88124be938" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374033 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="920c325a-f36b-4162-9d37-ea88124be938" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374047 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7af03157-ee92-4e72-a775-acaeabb73e65" containerName="marketplace-operator" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374055 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="7af03157-ee92-4e72-a775-acaeabb73e65" containerName="marketplace-operator" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374064 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="920c325a-f36b-4162-9d37-ea88124be938" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374072 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="920c325a-f36b-4162-9d37-ea88124be938" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374083 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b37a2847-a94a-4a0c-b092-1ed7155a2d35" containerName="controller-manager" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374091 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="b37a2847-a94a-4a0c-b092-1ed7155a2d35" containerName="controller-manager" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374102 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374109 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374120 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="891b78f7-509c-4e8d-b846-52881396a64d" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374127 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="891b78f7-509c-4e8d-b846-52881396a64d" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374140 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374148 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374157 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374165 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="extract-content" Jan 24 00:08:50 crc kubenswrapper[4676]: E0124 00:08:50.374173 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374181 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerName="extract-utilities" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374278 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4d7c63-cff0-4408-9cb6-450f3ebc53dd" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374290 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="b37a2847-a94a-4a0c-b092-1ed7155a2d35" containerName="controller-manager" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374303 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="920c325a-f36b-4162-9d37-ea88124be938" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374316 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="7af03157-ee92-4e72-a775-acaeabb73e65" containerName="marketplace-operator" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374325 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="891b78f7-509c-4e8d-b846-52881396a64d" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374339 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="10798203-b391-4a87-98a7-b41db2bbb0e2" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374350 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe4a6a13-8de6-4b4b-8bd7-3e2755e7cf93" containerName="route-controller-manager" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374360 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="592859d8-1f7e-4e35-acc9-635e130ad2d2" containerName="registry-server" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.374763 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.398298 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9qdxp"] Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.473325 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/698d0139-4397-4dd2-bdc1-580717c2f177-bound-sa-token\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.473666 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/698d0139-4397-4dd2-bdc1-580717c2f177-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.473797 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvllg\" (UniqueName: \"kubernetes.io/projected/698d0139-4397-4dd2-bdc1-580717c2f177-kube-api-access-pvllg\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.473904 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/698d0139-4397-4dd2-bdc1-580717c2f177-registry-certificates\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.474006 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/698d0139-4397-4dd2-bdc1-580717c2f177-trusted-ca\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.474170 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.474321 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/698d0139-4397-4dd2-bdc1-580717c2f177-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.474496 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/698d0139-4397-4dd2-bdc1-580717c2f177-registry-tls\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.500306 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.575462 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/698d0139-4397-4dd2-bdc1-580717c2f177-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.575506 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/698d0139-4397-4dd2-bdc1-580717c2f177-registry-tls\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.575532 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/698d0139-4397-4dd2-bdc1-580717c2f177-bound-sa-token\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.575566 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/698d0139-4397-4dd2-bdc1-580717c2f177-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.575590 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvllg\" (UniqueName: \"kubernetes.io/projected/698d0139-4397-4dd2-bdc1-580717c2f177-kube-api-access-pvllg\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.575612 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/698d0139-4397-4dd2-bdc1-580717c2f177-registry-certificates\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.575633 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/698d0139-4397-4dd2-bdc1-580717c2f177-trusted-ca\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.576511 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/698d0139-4397-4dd2-bdc1-580717c2f177-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.576904 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/698d0139-4397-4dd2-bdc1-580717c2f177-trusted-ca\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.577885 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/698d0139-4397-4dd2-bdc1-580717c2f177-registry-certificates\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.579747 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/698d0139-4397-4dd2-bdc1-580717c2f177-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.581626 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/698d0139-4397-4dd2-bdc1-580717c2f177-registry-tls\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.594154 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/698d0139-4397-4dd2-bdc1-580717c2f177-bound-sa-token\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.594590 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvllg\" (UniqueName: \"kubernetes.io/projected/698d0139-4397-4dd2-bdc1-580717c2f177-kube-api-access-pvllg\") pod \"image-registry-66df7c8f76-9qdxp\" (UID: \"698d0139-4397-4dd2-bdc1-580717c2f177\") " pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.686707 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7"] Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.687367 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.688554 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.693509 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.695248 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.695639 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.696022 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.696339 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.702098 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.713010 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6567888b4c-v7f69"] Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.713795 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.720454 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.720707 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.721937 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.722107 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.722305 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.724157 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7"] Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.724459 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.728238 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.747179 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6567888b4c-v7f69"] Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.778032 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-config\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.778085 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8gd6z\" (UniqueName: \"kubernetes.io/projected/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-kube-api-access-8gd6z\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.778112 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-serving-cert\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.778168 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-client-ca\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.875277 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9qdxp"] Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.879521 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5hnq\" (UniqueName: \"kubernetes.io/projected/3166691a-d300-4e14-a2ae-745fac50b3b4-kube-api-access-v5hnq\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.879565 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3166691a-d300-4e14-a2ae-745fac50b3b4-serving-cert\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.879593 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-client-ca\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.879754 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-client-ca\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.879856 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-proxy-ca-bundles\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.879913 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-config\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.879937 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-config\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.879973 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8gd6z\" (UniqueName: \"kubernetes.io/projected/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-kube-api-access-8gd6z\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.880007 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-serving-cert\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.881207 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-client-ca\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.881415 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-config\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.894108 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-serving-cert\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.897117 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8gd6z\" (UniqueName: \"kubernetes.io/projected/b7dc4ada-dd45-47fd-922a-903ba16f5b6f-kube-api-access-8gd6z\") pod \"route-controller-manager-5cd57b68f5-8z6p7\" (UID: \"b7dc4ada-dd45-47fd-922a-903ba16f5b6f\") " pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.981518 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-proxy-ca-bundles\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.981571 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-config\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.981607 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3166691a-d300-4e14-a2ae-745fac50b3b4-serving-cert\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.981624 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5hnq\" (UniqueName: \"kubernetes.io/projected/3166691a-d300-4e14-a2ae-745fac50b3b4-kube-api-access-v5hnq\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.981646 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-client-ca\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.982444 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-client-ca\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.983534 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-proxy-ca-bundles\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.983601 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-config\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.986537 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3166691a-d300-4e14-a2ae-745fac50b3b4-serving-cert\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:50 crc kubenswrapper[4676]: I0124 00:08:50.998094 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5hnq\" (UniqueName: \"kubernetes.io/projected/3166691a-d300-4e14-a2ae-745fac50b3b4-kube-api-access-v5hnq\") pod \"controller-manager-6567888b4c-v7f69\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.001002 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.037831 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.065398 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" event={"ID":"698d0139-4397-4dd2-bdc1-580717c2f177","Type":"ContainerStarted","Data":"92940ffc0d2cfca482946fb554a79d9370861525d9d58ba8db4994a3dc9d39cd"} Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.225393 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6567888b4c-v7f69"] Jan 24 00:08:51 crc kubenswrapper[4676]: W0124 00:08:51.230634 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3166691a_d300_4e14_a2ae_745fac50b3b4.slice/crio-37eb479d7ed3ff133df52eba62ac268cc9d612b8b14b6586faeb23724b08b838 WatchSource:0}: Error finding container 37eb479d7ed3ff133df52eba62ac268cc9d612b8b14b6586faeb23724b08b838: Status 404 returned error can't find the container with id 37eb479d7ed3ff133df52eba62ac268cc9d612b8b14b6586faeb23724b08b838 Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.384795 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7"] Jan 24 00:08:51 crc kubenswrapper[4676]: W0124 00:08:51.392352 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7dc4ada_dd45_47fd_922a_903ba16f5b6f.slice/crio-e066f82670a61a77c62917ec31f759a4319de1c12e3ea9dbf6cbe72af7361e7f WatchSource:0}: Error finding container e066f82670a61a77c62917ec31f759a4319de1c12e3ea9dbf6cbe72af7361e7f: Status 404 returned error can't find the container with id e066f82670a61a77c62917ec31f759a4319de1c12e3ea9dbf6cbe72af7361e7f Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.662498 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-h6lxx"] Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.664541 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.670952 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.691830 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h6lxx"] Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.792868 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdlrw\" (UniqueName: \"kubernetes.io/projected/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-kube-api-access-wdlrw\") pod \"certified-operators-h6lxx\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.792921 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-utilities\") pod \"certified-operators-h6lxx\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.792957 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-catalog-content\") pod \"certified-operators-h6lxx\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.893824 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-catalog-content\") pod \"certified-operators-h6lxx\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.894108 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdlrw\" (UniqueName: \"kubernetes.io/projected/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-kube-api-access-wdlrw\") pod \"certified-operators-h6lxx\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.894135 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-utilities\") pod \"certified-operators-h6lxx\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.894752 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-utilities\") pod \"certified-operators-h6lxx\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.894908 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-catalog-content\") pod \"certified-operators-h6lxx\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.919017 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdlrw\" (UniqueName: \"kubernetes.io/projected/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-kube-api-access-wdlrw\") pod \"certified-operators-h6lxx\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:51 crc kubenswrapper[4676]: I0124 00:08:51.997900 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.071211 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" event={"ID":"698d0139-4397-4dd2-bdc1-580717c2f177","Type":"ContainerStarted","Data":"454a2b3261dc96d1aa000c62b3271f6a19ae8d899e8d8d4d52f576846cb675e2"} Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.072100 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.073423 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" event={"ID":"3166691a-d300-4e14-a2ae-745fac50b3b4","Type":"ContainerStarted","Data":"17caedb43a800482408b315f9631ff9a96ef55fffba165823c14cc78155a7eef"} Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.073444 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" event={"ID":"3166691a-d300-4e14-a2ae-745fac50b3b4","Type":"ContainerStarted","Data":"37eb479d7ed3ff133df52eba62ac268cc9d612b8b14b6586faeb23724b08b838"} Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.073939 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.074768 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" event={"ID":"b7dc4ada-dd45-47fd-922a-903ba16f5b6f","Type":"ContainerStarted","Data":"4b28313a5d30610e5a876388bca48312314c883ec5de82b84586937fa888ecd3"} Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.074787 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" event={"ID":"b7dc4ada-dd45-47fd-922a-903ba16f5b6f","Type":"ContainerStarted","Data":"e066f82670a61a77c62917ec31f759a4319de1c12e3ea9dbf6cbe72af7361e7f"} Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.075184 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.087861 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.105558 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" podStartSLOduration=2.105543142 podStartE2EDuration="2.105543142s" podCreationTimestamp="2026-01-24 00:08:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:08:52.102486311 +0000 UTC m=+316.132457312" watchObservedRunningTime="2026-01-24 00:08:52.105543142 +0000 UTC m=+316.135514133" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.206729 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" podStartSLOduration=4.206714039 podStartE2EDuration="4.206714039s" podCreationTimestamp="2026-01-24 00:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:08:52.201276681 +0000 UTC m=+316.231247682" watchObservedRunningTime="2026-01-24 00:08:52.206714039 +0000 UTC m=+316.236685040" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.236472 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" podStartSLOduration=4.236450227 podStartE2EDuration="4.236450227s" podCreationTimestamp="2026-01-24 00:08:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:08:52.234643048 +0000 UTC m=+316.264614049" watchObservedRunningTime="2026-01-24 00:08:52.236450227 +0000 UTC m=+316.266421228" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.271563 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rjjnx"] Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.272519 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.284587 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.324712 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjjnx"] Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.363584 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5cd57b68f5-8z6p7" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.388349 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-h6lxx"] Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.406457 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/646456dc-35bc-4df2-8f92-55cdfefc6010-catalog-content\") pod \"community-operators-rjjnx\" (UID: \"646456dc-35bc-4df2-8f92-55cdfefc6010\") " pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.406521 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/646456dc-35bc-4df2-8f92-55cdfefc6010-utilities\") pod \"community-operators-rjjnx\" (UID: \"646456dc-35bc-4df2-8f92-55cdfefc6010\") " pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.406557 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5wp4\" (UniqueName: \"kubernetes.io/projected/646456dc-35bc-4df2-8f92-55cdfefc6010-kube-api-access-n5wp4\") pod \"community-operators-rjjnx\" (UID: \"646456dc-35bc-4df2-8f92-55cdfefc6010\") " pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.507388 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/646456dc-35bc-4df2-8f92-55cdfefc6010-utilities\") pod \"community-operators-rjjnx\" (UID: \"646456dc-35bc-4df2-8f92-55cdfefc6010\") " pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.507457 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5wp4\" (UniqueName: \"kubernetes.io/projected/646456dc-35bc-4df2-8f92-55cdfefc6010-kube-api-access-n5wp4\") pod \"community-operators-rjjnx\" (UID: \"646456dc-35bc-4df2-8f92-55cdfefc6010\") " pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.507514 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/646456dc-35bc-4df2-8f92-55cdfefc6010-catalog-content\") pod \"community-operators-rjjnx\" (UID: \"646456dc-35bc-4df2-8f92-55cdfefc6010\") " pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.507958 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/646456dc-35bc-4df2-8f92-55cdfefc6010-utilities\") pod \"community-operators-rjjnx\" (UID: \"646456dc-35bc-4df2-8f92-55cdfefc6010\") " pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.508573 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/646456dc-35bc-4df2-8f92-55cdfefc6010-catalog-content\") pod \"community-operators-rjjnx\" (UID: \"646456dc-35bc-4df2-8f92-55cdfefc6010\") " pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.524742 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5wp4\" (UniqueName: \"kubernetes.io/projected/646456dc-35bc-4df2-8f92-55cdfefc6010-kube-api-access-n5wp4\") pod \"community-operators-rjjnx\" (UID: \"646456dc-35bc-4df2-8f92-55cdfefc6010\") " pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.592931 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:08:52 crc kubenswrapper[4676]: I0124 00:08:52.993684 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjjnx"] Jan 24 00:08:52 crc kubenswrapper[4676]: W0124 00:08:52.998999 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod646456dc_35bc_4df2_8f92_55cdfefc6010.slice/crio-873d12d25a978b1d381c40e12ad2f6a27ab8fbf2fce4ad20e07e4752dfe1ed65 WatchSource:0}: Error finding container 873d12d25a978b1d381c40e12ad2f6a27ab8fbf2fce4ad20e07e4752dfe1ed65: Status 404 returned error can't find the container with id 873d12d25a978b1d381c40e12ad2f6a27ab8fbf2fce4ad20e07e4752dfe1ed65 Jan 24 00:08:53 crc kubenswrapper[4676]: I0124 00:08:53.081109 4676 generic.go:334] "Generic (PLEG): container finished" podID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerID="e1126bba560d6ad086386fdbfde6ce098ca6ec45eb8ef89aa5c2a014e09ba3c2" exitCode=0 Jan 24 00:08:53 crc kubenswrapper[4676]: I0124 00:08:53.081170 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6lxx" event={"ID":"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec","Type":"ContainerDied","Data":"e1126bba560d6ad086386fdbfde6ce098ca6ec45eb8ef89aa5c2a014e09ba3c2"} Jan 24 00:08:53 crc kubenswrapper[4676]: I0124 00:08:53.081201 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6lxx" event={"ID":"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec","Type":"ContainerStarted","Data":"fb5783295da5666c028415e55625575a6eddcc0d1dfa67d4345fe77ee0fb6e6b"} Jan 24 00:08:53 crc kubenswrapper[4676]: I0124 00:08:53.082598 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjjnx" event={"ID":"646456dc-35bc-4df2-8f92-55cdfefc6010","Type":"ContainerStarted","Data":"873d12d25a978b1d381c40e12ad2f6a27ab8fbf2fce4ad20e07e4752dfe1ed65"} Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.058050 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vw7tg"] Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.059163 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.061158 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.072402 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vw7tg"] Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.090080 4676 generic.go:334] "Generic (PLEG): container finished" podID="646456dc-35bc-4df2-8f92-55cdfefc6010" containerID="58c7ef0829c97d66dbcc2b639b255fe055d395775dc8bbd554aec915453da312" exitCode=0 Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.090151 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjjnx" event={"ID":"646456dc-35bc-4df2-8f92-55cdfefc6010","Type":"ContainerDied","Data":"58c7ef0829c97d66dbcc2b639b255fe055d395775dc8bbd554aec915453da312"} Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.133534 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c82da49-780b-431e-bfe7-d52ce3bcb623-catalog-content\") pod \"redhat-marketplace-vw7tg\" (UID: \"4c82da49-780b-431e-bfe7-d52ce3bcb623\") " pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.133687 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rmtt\" (UniqueName: \"kubernetes.io/projected/4c82da49-780b-431e-bfe7-d52ce3bcb623-kube-api-access-8rmtt\") pod \"redhat-marketplace-vw7tg\" (UID: \"4c82da49-780b-431e-bfe7-d52ce3bcb623\") " pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.133825 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c82da49-780b-431e-bfe7-d52ce3bcb623-utilities\") pod \"redhat-marketplace-vw7tg\" (UID: \"4c82da49-780b-431e-bfe7-d52ce3bcb623\") " pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.234563 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c82da49-780b-431e-bfe7-d52ce3bcb623-catalog-content\") pod \"redhat-marketplace-vw7tg\" (UID: \"4c82da49-780b-431e-bfe7-d52ce3bcb623\") " pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.234682 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rmtt\" (UniqueName: \"kubernetes.io/projected/4c82da49-780b-431e-bfe7-d52ce3bcb623-kube-api-access-8rmtt\") pod \"redhat-marketplace-vw7tg\" (UID: \"4c82da49-780b-431e-bfe7-d52ce3bcb623\") " pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.234725 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c82da49-780b-431e-bfe7-d52ce3bcb623-utilities\") pod \"redhat-marketplace-vw7tg\" (UID: \"4c82da49-780b-431e-bfe7-d52ce3bcb623\") " pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.235101 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4c82da49-780b-431e-bfe7-d52ce3bcb623-utilities\") pod \"redhat-marketplace-vw7tg\" (UID: \"4c82da49-780b-431e-bfe7-d52ce3bcb623\") " pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.235311 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4c82da49-780b-431e-bfe7-d52ce3bcb623-catalog-content\") pod \"redhat-marketplace-vw7tg\" (UID: \"4c82da49-780b-431e-bfe7-d52ce3bcb623\") " pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.258721 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rmtt\" (UniqueName: \"kubernetes.io/projected/4c82da49-780b-431e-bfe7-d52ce3bcb623-kube-api-access-8rmtt\") pod \"redhat-marketplace-vw7tg\" (UID: \"4c82da49-780b-431e-bfe7-d52ce3bcb623\") " pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.375174 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.655583 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6mz4x"] Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.656752 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.659693 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.672731 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6mz4x"] Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.741935 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr7r2\" (UniqueName: \"kubernetes.io/projected/829774a4-dd19-462e-829c-f201bddf6886-kube-api-access-fr7r2\") pod \"redhat-operators-6mz4x\" (UID: \"829774a4-dd19-462e-829c-f201bddf6886\") " pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.742228 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/829774a4-dd19-462e-829c-f201bddf6886-catalog-content\") pod \"redhat-operators-6mz4x\" (UID: \"829774a4-dd19-462e-829c-f201bddf6886\") " pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.742347 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/829774a4-dd19-462e-829c-f201bddf6886-utilities\") pod \"redhat-operators-6mz4x\" (UID: \"829774a4-dd19-462e-829c-f201bddf6886\") " pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.777664 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vw7tg"] Jan 24 00:08:54 crc kubenswrapper[4676]: W0124 00:08:54.786761 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c82da49_780b_431e_bfe7_d52ce3bcb623.slice/crio-87779d34b0d4df9f6fb9a6f75cdce41fa7658d4a74055ee1cd77144d2add7d95 WatchSource:0}: Error finding container 87779d34b0d4df9f6fb9a6f75cdce41fa7658d4a74055ee1cd77144d2add7d95: Status 404 returned error can't find the container with id 87779d34b0d4df9f6fb9a6f75cdce41fa7658d4a74055ee1cd77144d2add7d95 Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.843770 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr7r2\" (UniqueName: \"kubernetes.io/projected/829774a4-dd19-462e-829c-f201bddf6886-kube-api-access-fr7r2\") pod \"redhat-operators-6mz4x\" (UID: \"829774a4-dd19-462e-829c-f201bddf6886\") " pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.844030 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/829774a4-dd19-462e-829c-f201bddf6886-catalog-content\") pod \"redhat-operators-6mz4x\" (UID: \"829774a4-dd19-462e-829c-f201bddf6886\") " pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.844196 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/829774a4-dd19-462e-829c-f201bddf6886-utilities\") pod \"redhat-operators-6mz4x\" (UID: \"829774a4-dd19-462e-829c-f201bddf6886\") " pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.844576 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/829774a4-dd19-462e-829c-f201bddf6886-catalog-content\") pod \"redhat-operators-6mz4x\" (UID: \"829774a4-dd19-462e-829c-f201bddf6886\") " pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.844679 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/829774a4-dd19-462e-829c-f201bddf6886-utilities\") pod \"redhat-operators-6mz4x\" (UID: \"829774a4-dd19-462e-829c-f201bddf6886\") " pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.869717 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr7r2\" (UniqueName: \"kubernetes.io/projected/829774a4-dd19-462e-829c-f201bddf6886-kube-api-access-fr7r2\") pod \"redhat-operators-6mz4x\" (UID: \"829774a4-dd19-462e-829c-f201bddf6886\") " pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:54 crc kubenswrapper[4676]: I0124 00:08:54.972066 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:08:55 crc kubenswrapper[4676]: I0124 00:08:55.100663 4676 generic.go:334] "Generic (PLEG): container finished" podID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerID="52bc77f8ab557f140d976553f763b957c850be181430d480625497a667852cf5" exitCode=0 Jan 24 00:08:55 crc kubenswrapper[4676]: I0124 00:08:55.100873 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6lxx" event={"ID":"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec","Type":"ContainerDied","Data":"52bc77f8ab557f140d976553f763b957c850be181430d480625497a667852cf5"} Jan 24 00:08:55 crc kubenswrapper[4676]: I0124 00:08:55.113974 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vw7tg" event={"ID":"4c82da49-780b-431e-bfe7-d52ce3bcb623","Type":"ContainerStarted","Data":"87779d34b0d4df9f6fb9a6f75cdce41fa7658d4a74055ee1cd77144d2add7d95"} Jan 24 00:08:55 crc kubenswrapper[4676]: I0124 00:08:55.378047 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6mz4x"] Jan 24 00:08:55 crc kubenswrapper[4676]: W0124 00:08:55.388639 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod829774a4_dd19_462e_829c_f201bddf6886.slice/crio-d40c66edb9d526f7eafa7bcd62806d639f76eae15baf3efd412242c5e14b1ac7 WatchSource:0}: Error finding container d40c66edb9d526f7eafa7bcd62806d639f76eae15baf3efd412242c5e14b1ac7: Status 404 returned error can't find the container with id d40c66edb9d526f7eafa7bcd62806d639f76eae15baf3efd412242c5e14b1ac7 Jan 24 00:08:56 crc kubenswrapper[4676]: I0124 00:08:56.120707 4676 generic.go:334] "Generic (PLEG): container finished" podID="829774a4-dd19-462e-829c-f201bddf6886" containerID="b4fa1a55d653742673b269e36abd70a8ffdf13c32586d0ad8c05867c492d81d6" exitCode=0 Jan 24 00:08:56 crc kubenswrapper[4676]: I0124 00:08:56.120820 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mz4x" event={"ID":"829774a4-dd19-462e-829c-f201bddf6886","Type":"ContainerDied","Data":"b4fa1a55d653742673b269e36abd70a8ffdf13c32586d0ad8c05867c492d81d6"} Jan 24 00:08:56 crc kubenswrapper[4676]: I0124 00:08:56.122406 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mz4x" event={"ID":"829774a4-dd19-462e-829c-f201bddf6886","Type":"ContainerStarted","Data":"d40c66edb9d526f7eafa7bcd62806d639f76eae15baf3efd412242c5e14b1ac7"} Jan 24 00:08:56 crc kubenswrapper[4676]: I0124 00:08:56.127431 4676 generic.go:334] "Generic (PLEG): container finished" podID="4c82da49-780b-431e-bfe7-d52ce3bcb623" containerID="2ad57c4829ee7e5ba8d86b54b074c291baf42d2492d0a4ade40e4ddd0a9e2c36" exitCode=0 Jan 24 00:08:56 crc kubenswrapper[4676]: I0124 00:08:56.127485 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vw7tg" event={"ID":"4c82da49-780b-431e-bfe7-d52ce3bcb623","Type":"ContainerDied","Data":"2ad57c4829ee7e5ba8d86b54b074c291baf42d2492d0a4ade40e4ddd0a9e2c36"} Jan 24 00:08:57 crc kubenswrapper[4676]: I0124 00:08:57.135021 4676 generic.go:334] "Generic (PLEG): container finished" podID="646456dc-35bc-4df2-8f92-55cdfefc6010" containerID="6d56dc187e7b9a09f0626f71f550fffbc8a65054f1ea765e965463593c59a4f6" exitCode=0 Jan 24 00:08:57 crc kubenswrapper[4676]: I0124 00:08:57.135111 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjjnx" event={"ID":"646456dc-35bc-4df2-8f92-55cdfefc6010","Type":"ContainerDied","Data":"6d56dc187e7b9a09f0626f71f550fffbc8a65054f1ea765e965463593c59a4f6"} Jan 24 00:08:57 crc kubenswrapper[4676]: I0124 00:08:57.141573 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6lxx" event={"ID":"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec","Type":"ContainerStarted","Data":"08fb9c82277ed77127a24cba780debc02f8a59da1333b461232dac0199f577d3"} Jan 24 00:08:58 crc kubenswrapper[4676]: I0124 00:08:58.147925 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vw7tg" event={"ID":"4c82da49-780b-431e-bfe7-d52ce3bcb623","Type":"ContainerStarted","Data":"b735d13a2378d0ba980128541eea2a3c9ffff1e2f4f3b214c7495548bc254a78"} Jan 24 00:08:58 crc kubenswrapper[4676]: I0124 00:08:58.150903 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mz4x" event={"ID":"829774a4-dd19-462e-829c-f201bddf6886","Type":"ContainerStarted","Data":"0eeb0c466ea082c5a0617505ab4895dc30f2bf85182ce4f980c155a9b92eaf6a"} Jan 24 00:08:59 crc kubenswrapper[4676]: I0124 00:08:59.159310 4676 generic.go:334] "Generic (PLEG): container finished" podID="829774a4-dd19-462e-829c-f201bddf6886" containerID="0eeb0c466ea082c5a0617505ab4895dc30f2bf85182ce4f980c155a9b92eaf6a" exitCode=0 Jan 24 00:08:59 crc kubenswrapper[4676]: I0124 00:08:59.159690 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mz4x" event={"ID":"829774a4-dd19-462e-829c-f201bddf6886","Type":"ContainerDied","Data":"0eeb0c466ea082c5a0617505ab4895dc30f2bf85182ce4f980c155a9b92eaf6a"} Jan 24 00:08:59 crc kubenswrapper[4676]: I0124 00:08:59.161799 4676 generic.go:334] "Generic (PLEG): container finished" podID="4c82da49-780b-431e-bfe7-d52ce3bcb623" containerID="b735d13a2378d0ba980128541eea2a3c9ffff1e2f4f3b214c7495548bc254a78" exitCode=0 Jan 24 00:08:59 crc kubenswrapper[4676]: I0124 00:08:59.161911 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vw7tg" event={"ID":"4c82da49-780b-431e-bfe7-d52ce3bcb623","Type":"ContainerDied","Data":"b735d13a2378d0ba980128541eea2a3c9ffff1e2f4f3b214c7495548bc254a78"} Jan 24 00:08:59 crc kubenswrapper[4676]: I0124 00:08:59.168931 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjjnx" event={"ID":"646456dc-35bc-4df2-8f92-55cdfefc6010","Type":"ContainerStarted","Data":"bce9f6b37e559b5c43b32d751e038d3ab9cf0038b0ce24cfb67169da92057021"} Jan 24 00:08:59 crc kubenswrapper[4676]: I0124 00:08:59.187436 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-h6lxx" podStartSLOduration=4.781868617 podStartE2EDuration="8.187416085s" podCreationTimestamp="2026-01-24 00:08:51 +0000 UTC" firstStartedPulling="2026-01-24 00:08:53.084167799 +0000 UTC m=+317.114138840" lastFinishedPulling="2026-01-24 00:08:56.489715307 +0000 UTC m=+320.519686308" observedRunningTime="2026-01-24 00:08:58.21250774 +0000 UTC m=+322.242478741" watchObservedRunningTime="2026-01-24 00:08:59.187416085 +0000 UTC m=+323.217387096" Jan 24 00:08:59 crc kubenswrapper[4676]: I0124 00:08:59.205596 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rjjnx" podStartSLOduration=3.10911072 podStartE2EDuration="7.205580802s" podCreationTimestamp="2026-01-24 00:08:52 +0000 UTC" firstStartedPulling="2026-01-24 00:08:54.092863375 +0000 UTC m=+318.122834376" lastFinishedPulling="2026-01-24 00:08:58.189333457 +0000 UTC m=+322.219304458" observedRunningTime="2026-01-24 00:08:59.204965951 +0000 UTC m=+323.234936972" watchObservedRunningTime="2026-01-24 00:08:59.205580802 +0000 UTC m=+323.235551803" Jan 24 00:09:02 crc kubenswrapper[4676]: I0124 00:09:01.999955 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:09:02 crc kubenswrapper[4676]: I0124 00:09:02.003490 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:09:02 crc kubenswrapper[4676]: I0124 00:09:02.118518 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:09:02 crc kubenswrapper[4676]: I0124 00:09:02.209698 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vw7tg" event={"ID":"4c82da49-780b-431e-bfe7-d52ce3bcb623","Type":"ContainerStarted","Data":"557ff3d4bd760ab14450e8412163fa653f384dce00b97bc2753f06d8fb200fb8"} Jan 24 00:09:02 crc kubenswrapper[4676]: I0124 00:09:02.266525 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:09:02 crc kubenswrapper[4676]: I0124 00:09:02.593442 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:09:02 crc kubenswrapper[4676]: I0124 00:09:02.593508 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:09:02 crc kubenswrapper[4676]: I0124 00:09:02.651113 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:09:03 crc kubenswrapper[4676]: I0124 00:09:03.216427 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6mz4x" event={"ID":"829774a4-dd19-462e-829c-f201bddf6886","Type":"ContainerStarted","Data":"1f4517eae223160755fe67ca42667d4b94c3389329951409164d7c368755eb3d"} Jan 24 00:09:03 crc kubenswrapper[4676]: I0124 00:09:03.234048 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vw7tg" podStartSLOduration=3.403564113 podStartE2EDuration="9.234033197s" podCreationTimestamp="2026-01-24 00:08:54 +0000 UTC" firstStartedPulling="2026-01-24 00:08:56.144401959 +0000 UTC m=+320.174372960" lastFinishedPulling="2026-01-24 00:09:01.974871003 +0000 UTC m=+326.004842044" observedRunningTime="2026-01-24 00:09:03.233512979 +0000 UTC m=+327.263484000" watchObservedRunningTime="2026-01-24 00:09:03.234033197 +0000 UTC m=+327.264004198" Jan 24 00:09:03 crc kubenswrapper[4676]: I0124 00:09:03.249345 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6mz4x" podStartSLOduration=3.405570291 podStartE2EDuration="9.24932428s" podCreationTimestamp="2026-01-24 00:08:54 +0000 UTC" firstStartedPulling="2026-01-24 00:08:56.124143714 +0000 UTC m=+320.154114735" lastFinishedPulling="2026-01-24 00:09:01.967897723 +0000 UTC m=+325.997868724" observedRunningTime="2026-01-24 00:09:03.247639614 +0000 UTC m=+327.277610625" watchObservedRunningTime="2026-01-24 00:09:03.24932428 +0000 UTC m=+327.279295291" Jan 24 00:09:03 crc kubenswrapper[4676]: I0124 00:09:03.262020 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rjjnx" Jan 24 00:09:04 crc kubenswrapper[4676]: I0124 00:09:04.375557 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:09:04 crc kubenswrapper[4676]: I0124 00:09:04.375630 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:09:04 crc kubenswrapper[4676]: I0124 00:09:04.436959 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:09:04 crc kubenswrapper[4676]: I0124 00:09:04.973028 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:09:04 crc kubenswrapper[4676]: I0124 00:09:04.973585 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:09:06 crc kubenswrapper[4676]: I0124 00:09:06.011060 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6mz4x" podUID="829774a4-dd19-462e-829c-f201bddf6886" containerName="registry-server" probeResult="failure" output=< Jan 24 00:09:06 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:09:06 crc kubenswrapper[4676]: > Jan 24 00:09:10 crc kubenswrapper[4676]: I0124 00:09:10.701276 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-9qdxp" Jan 24 00:09:10 crc kubenswrapper[4676]: I0124 00:09:10.782556 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shf2w"] Jan 24 00:09:14 crc kubenswrapper[4676]: I0124 00:09:14.432968 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vw7tg" Jan 24 00:09:15 crc kubenswrapper[4676]: I0124 00:09:15.033789 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:09:15 crc kubenswrapper[4676]: I0124 00:09:15.095449 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6mz4x" Jan 24 00:09:35 crc kubenswrapper[4676]: I0124 00:09:35.847606 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" podUID="9887557b-81eb-4651-8da2-fd34d7b0be97" containerName="registry" containerID="cri-o://38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472" gracePeriod=30 Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.224259 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.408288 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-certificates\") pod \"9887557b-81eb-4651-8da2-fd34d7b0be97\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.408343 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9887557b-81eb-4651-8da2-fd34d7b0be97-ca-trust-extracted\") pod \"9887557b-81eb-4651-8da2-fd34d7b0be97\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.408419 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-trusted-ca\") pod \"9887557b-81eb-4651-8da2-fd34d7b0be97\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.408577 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9887557b-81eb-4651-8da2-fd34d7b0be97\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.408643 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlknz\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-kube-api-access-hlknz\") pod \"9887557b-81eb-4651-8da2-fd34d7b0be97\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.408683 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9887557b-81eb-4651-8da2-fd34d7b0be97-installation-pull-secrets\") pod \"9887557b-81eb-4651-8da2-fd34d7b0be97\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.408729 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-tls\") pod \"9887557b-81eb-4651-8da2-fd34d7b0be97\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.408755 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-bound-sa-token\") pod \"9887557b-81eb-4651-8da2-fd34d7b0be97\" (UID: \"9887557b-81eb-4651-8da2-fd34d7b0be97\") " Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.409274 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9887557b-81eb-4651-8da2-fd34d7b0be97" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.410007 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9887557b-81eb-4651-8da2-fd34d7b0be97" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.417809 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-kube-api-access-hlknz" (OuterVolumeSpecName: "kube-api-access-hlknz") pod "9887557b-81eb-4651-8da2-fd34d7b0be97" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97"). InnerVolumeSpecName "kube-api-access-hlknz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.417916 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9887557b-81eb-4651-8da2-fd34d7b0be97-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9887557b-81eb-4651-8da2-fd34d7b0be97" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.418034 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9887557b-81eb-4651-8da2-fd34d7b0be97" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.423047 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9887557b-81eb-4651-8da2-fd34d7b0be97" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.423872 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9887557b-81eb-4651-8da2-fd34d7b0be97" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.437407 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9887557b-81eb-4651-8da2-fd34d7b0be97-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9887557b-81eb-4651-8da2-fd34d7b0be97" (UID: "9887557b-81eb-4651-8da2-fd34d7b0be97"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.450744 4676 generic.go:334] "Generic (PLEG): container finished" podID="9887557b-81eb-4651-8da2-fd34d7b0be97" containerID="38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472" exitCode=0 Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.450826 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.450856 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" event={"ID":"9887557b-81eb-4651-8da2-fd34d7b0be97","Type":"ContainerDied","Data":"38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472"} Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.451191 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-shf2w" event={"ID":"9887557b-81eb-4651-8da2-fd34d7b0be97","Type":"ContainerDied","Data":"927b1ecf53efd6bf7705450ed6ba3d508cc6d33832a7f941121086cc971d4adb"} Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.451218 4676 scope.go:117] "RemoveContainer" containerID="38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.484313 4676 scope.go:117] "RemoveContainer" containerID="38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.484821 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shf2w"] Jan 24 00:09:36 crc kubenswrapper[4676]: E0124 00:09:36.485192 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472\": container with ID starting with 38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472 not found: ID does not exist" containerID="38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.485239 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472"} err="failed to get container status \"38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472\": rpc error: code = NotFound desc = could not find container \"38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472\": container with ID starting with 38b74903d521a57e98f715c4b02dad7881bee2134fc7d9ceec4368f1ee47f472 not found: ID does not exist" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.488665 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shf2w"] Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.509684 4676 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.509713 4676 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.509723 4676 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.509733 4676 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9887557b-81eb-4651-8da2-fd34d7b0be97-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.509744 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9887557b-81eb-4651-8da2-fd34d7b0be97-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.509754 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlknz\" (UniqueName: \"kubernetes.io/projected/9887557b-81eb-4651-8da2-fd34d7b0be97-kube-api-access-hlknz\") on node \"crc\" DevicePath \"\"" Jan 24 00:09:36 crc kubenswrapper[4676]: I0124 00:09:36.509764 4676 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9887557b-81eb-4651-8da2-fd34d7b0be97-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 24 00:09:38 crc kubenswrapper[4676]: I0124 00:09:38.263988 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9887557b-81eb-4651-8da2-fd34d7b0be97" path="/var/lib/kubelet/pods/9887557b-81eb-4651-8da2-fd34d7b0be97/volumes" Jan 24 00:09:39 crc kubenswrapper[4676]: I0124 00:09:39.363972 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:09:39 crc kubenswrapper[4676]: I0124 00:09:39.364031 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.351958 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6567888b4c-v7f69"] Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.353046 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" podUID="3166691a-d300-4e14-a2ae-745fac50b3b4" containerName="controller-manager" containerID="cri-o://17caedb43a800482408b315f9631ff9a96ef55fffba165823c14cc78155a7eef" gracePeriod=30 Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.627822 4676 generic.go:334] "Generic (PLEG): container finished" podID="3166691a-d300-4e14-a2ae-745fac50b3b4" containerID="17caedb43a800482408b315f9631ff9a96ef55fffba165823c14cc78155a7eef" exitCode=0 Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.627893 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" event={"ID":"3166691a-d300-4e14-a2ae-745fac50b3b4","Type":"ContainerDied","Data":"17caedb43a800482408b315f9631ff9a96ef55fffba165823c14cc78155a7eef"} Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.718449 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.842478 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5hnq\" (UniqueName: \"kubernetes.io/projected/3166691a-d300-4e14-a2ae-745fac50b3b4-kube-api-access-v5hnq\") pod \"3166691a-d300-4e14-a2ae-745fac50b3b4\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.842572 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-config\") pod \"3166691a-d300-4e14-a2ae-745fac50b3b4\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.842619 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-proxy-ca-bundles\") pod \"3166691a-d300-4e14-a2ae-745fac50b3b4\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.842652 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3166691a-d300-4e14-a2ae-745fac50b3b4-serving-cert\") pod \"3166691a-d300-4e14-a2ae-745fac50b3b4\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.842686 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-client-ca\") pod \"3166691a-d300-4e14-a2ae-745fac50b3b4\" (UID: \"3166691a-d300-4e14-a2ae-745fac50b3b4\") " Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.843766 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-client-ca" (OuterVolumeSpecName: "client-ca") pod "3166691a-d300-4e14-a2ae-745fac50b3b4" (UID: "3166691a-d300-4e14-a2ae-745fac50b3b4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.843812 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3166691a-d300-4e14-a2ae-745fac50b3b4" (UID: "3166691a-d300-4e14-a2ae-745fac50b3b4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.846001 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-config" (OuterVolumeSpecName: "config") pod "3166691a-d300-4e14-a2ae-745fac50b3b4" (UID: "3166691a-d300-4e14-a2ae-745fac50b3b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.848042 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3166691a-d300-4e14-a2ae-745fac50b3b4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3166691a-d300-4e14-a2ae-745fac50b3b4" (UID: "3166691a-d300-4e14-a2ae-745fac50b3b4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.848513 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3166691a-d300-4e14-a2ae-745fac50b3b4-kube-api-access-v5hnq" (OuterVolumeSpecName: "kube-api-access-v5hnq") pod "3166691a-d300-4e14-a2ae-745fac50b3b4" (UID: "3166691a-d300-4e14-a2ae-745fac50b3b4"). InnerVolumeSpecName "kube-api-access-v5hnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.943775 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5hnq\" (UniqueName: \"kubernetes.io/projected/3166691a-d300-4e14-a2ae-745fac50b3b4-kube-api-access-v5hnq\") on node \"crc\" DevicePath \"\"" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.943809 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.943824 4676 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.943834 4676 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3166691a-d300-4e14-a2ae-745fac50b3b4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:10:07 crc kubenswrapper[4676]: I0124 00:10:07.943846 4676 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3166691a-d300-4e14-a2ae-745fac50b3b4-client-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.637506 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" event={"ID":"3166691a-d300-4e14-a2ae-745fac50b3b4","Type":"ContainerDied","Data":"37eb479d7ed3ff133df52eba62ac268cc9d612b8b14b6586faeb23724b08b838"} Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.637663 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6567888b4c-v7f69" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.639014 4676 scope.go:117] "RemoveContainer" containerID="17caedb43a800482408b315f9631ff9a96ef55fffba165823c14cc78155a7eef" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.661279 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6567888b4c-v7f69"] Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.665366 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6567888b4c-v7f69"] Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.742719 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-765455979b-gkrv4"] Jan 24 00:10:08 crc kubenswrapper[4676]: E0124 00:10:08.743108 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9887557b-81eb-4651-8da2-fd34d7b0be97" containerName="registry" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.743140 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9887557b-81eb-4651-8da2-fd34d7b0be97" containerName="registry" Jan 24 00:10:08 crc kubenswrapper[4676]: E0124 00:10:08.743171 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3166691a-d300-4e14-a2ae-745fac50b3b4" containerName="controller-manager" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.743186 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3166691a-d300-4e14-a2ae-745fac50b3b4" containerName="controller-manager" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.743362 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3166691a-d300-4e14-a2ae-745fac50b3b4" containerName="controller-manager" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.743423 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="9887557b-81eb-4651-8da2-fd34d7b0be97" containerName="registry" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.744007 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.746780 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.747044 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.747127 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.747281 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.747769 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.755183 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.758350 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-765455979b-gkrv4"] Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.775037 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.855350 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444c268d-a28d-4b7b-a1a3-6bbe0792653e-config\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.855433 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/444c268d-a28d-4b7b-a1a3-6bbe0792653e-client-ca\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.855469 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jkk4\" (UniqueName: \"kubernetes.io/projected/444c268d-a28d-4b7b-a1a3-6bbe0792653e-kube-api-access-8jkk4\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.855495 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444c268d-a28d-4b7b-a1a3-6bbe0792653e-serving-cert\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.855529 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/444c268d-a28d-4b7b-a1a3-6bbe0792653e-proxy-ca-bundles\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.957549 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444c268d-a28d-4b7b-a1a3-6bbe0792653e-config\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.957743 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/444c268d-a28d-4b7b-a1a3-6bbe0792653e-client-ca\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.957835 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jkk4\" (UniqueName: \"kubernetes.io/projected/444c268d-a28d-4b7b-a1a3-6bbe0792653e-kube-api-access-8jkk4\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.957896 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444c268d-a28d-4b7b-a1a3-6bbe0792653e-serving-cert\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.957957 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/444c268d-a28d-4b7b-a1a3-6bbe0792653e-proxy-ca-bundles\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.959994 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/444c268d-a28d-4b7b-a1a3-6bbe0792653e-proxy-ca-bundles\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.960992 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444c268d-a28d-4b7b-a1a3-6bbe0792653e-config\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.962322 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/444c268d-a28d-4b7b-a1a3-6bbe0792653e-client-ca\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.966124 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444c268d-a28d-4b7b-a1a3-6bbe0792653e-serving-cert\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:08 crc kubenswrapper[4676]: I0124 00:10:08.990518 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jkk4\" (UniqueName: \"kubernetes.io/projected/444c268d-a28d-4b7b-a1a3-6bbe0792653e-kube-api-access-8jkk4\") pod \"controller-manager-765455979b-gkrv4\" (UID: \"444c268d-a28d-4b7b-a1a3-6bbe0792653e\") " pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:09 crc kubenswrapper[4676]: I0124 00:10:09.077751 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:09 crc kubenswrapper[4676]: I0124 00:10:09.364203 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:10:09 crc kubenswrapper[4676]: I0124 00:10:09.364580 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:10:09 crc kubenswrapper[4676]: I0124 00:10:09.551424 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-765455979b-gkrv4"] Jan 24 00:10:09 crc kubenswrapper[4676]: I0124 00:10:09.646173 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" event={"ID":"444c268d-a28d-4b7b-a1a3-6bbe0792653e","Type":"ContainerStarted","Data":"699cf66899ea5853552f1ca6f0ef46e8661aad000af615b9c0e74305c336f797"} Jan 24 00:10:10 crc kubenswrapper[4676]: I0124 00:10:10.263633 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3166691a-d300-4e14-a2ae-745fac50b3b4" path="/var/lib/kubelet/pods/3166691a-d300-4e14-a2ae-745fac50b3b4/volumes" Jan 24 00:10:10 crc kubenswrapper[4676]: I0124 00:10:10.654469 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" event={"ID":"444c268d-a28d-4b7b-a1a3-6bbe0792653e","Type":"ContainerStarted","Data":"967d2fd4b40dab65607f024038c708140924dc24afa407932443cf8620caadf3"} Jan 24 00:10:10 crc kubenswrapper[4676]: I0124 00:10:10.655206 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:10 crc kubenswrapper[4676]: I0124 00:10:10.662277 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" Jan 24 00:10:10 crc kubenswrapper[4676]: I0124 00:10:10.682178 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-765455979b-gkrv4" podStartSLOduration=3.682139464 podStartE2EDuration="3.682139464s" podCreationTimestamp="2026-01-24 00:10:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:10:10.67422713 +0000 UTC m=+394.704198141" watchObservedRunningTime="2026-01-24 00:10:10.682139464 +0000 UTC m=+394.712110515" Jan 24 00:10:39 crc kubenswrapper[4676]: I0124 00:10:39.364803 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:10:39 crc kubenswrapper[4676]: I0124 00:10:39.365456 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:10:39 crc kubenswrapper[4676]: I0124 00:10:39.365527 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:10:39 crc kubenswrapper[4676]: I0124 00:10:39.366304 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6a2899a2bd48b07d11a54988ae27807805689a40bb41df1e769ca4da39298f4c"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:10:39 crc kubenswrapper[4676]: I0124 00:10:39.366420 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://6a2899a2bd48b07d11a54988ae27807805689a40bb41df1e769ca4da39298f4c" gracePeriod=600 Jan 24 00:10:39 crc kubenswrapper[4676]: I0124 00:10:39.848808 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="6a2899a2bd48b07d11a54988ae27807805689a40bb41df1e769ca4da39298f4c" exitCode=0 Jan 24 00:10:39 crc kubenswrapper[4676]: I0124 00:10:39.848928 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"6a2899a2bd48b07d11a54988ae27807805689a40bb41df1e769ca4da39298f4c"} Jan 24 00:10:39 crc kubenswrapper[4676]: I0124 00:10:39.849225 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"51936651fd255c79ecefd9c199d1a0336083a72c956e86d582874060fc907470"} Jan 24 00:10:39 crc kubenswrapper[4676]: I0124 00:10:39.849254 4676 scope.go:117] "RemoveContainer" containerID="9bf2fa5fb75b902d85e89d6ce3189bb1074a855a03752ec7f4fd03195945544d" Jan 24 00:12:39 crc kubenswrapper[4676]: I0124 00:12:39.364298 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:12:39 crc kubenswrapper[4676]: I0124 00:12:39.365595 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:13:09 crc kubenswrapper[4676]: I0124 00:13:09.365178 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:13:09 crc kubenswrapper[4676]: I0124 00:13:09.365842 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:13:39 crc kubenswrapper[4676]: I0124 00:13:39.364593 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:13:39 crc kubenswrapper[4676]: I0124 00:13:39.365244 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:13:39 crc kubenswrapper[4676]: I0124 00:13:39.365308 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:13:39 crc kubenswrapper[4676]: I0124 00:13:39.366344 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"51936651fd255c79ecefd9c199d1a0336083a72c956e86d582874060fc907470"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:13:39 crc kubenswrapper[4676]: I0124 00:13:39.366632 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://51936651fd255c79ecefd9c199d1a0336083a72c956e86d582874060fc907470" gracePeriod=600 Jan 24 00:13:40 crc kubenswrapper[4676]: I0124 00:13:40.165962 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="51936651fd255c79ecefd9c199d1a0336083a72c956e86d582874060fc907470" exitCode=0 Jan 24 00:13:40 crc kubenswrapper[4676]: I0124 00:13:40.166503 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"51936651fd255c79ecefd9c199d1a0336083a72c956e86d582874060fc907470"} Jan 24 00:13:40 crc kubenswrapper[4676]: I0124 00:13:40.166535 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"daf4dfd81dc7faee8c5a37cce872ffde5731f2d91708788dd42d2993fec18ba6"} Jan 24 00:13:40 crc kubenswrapper[4676]: I0124 00:13:40.166566 4676 scope.go:117] "RemoveContainer" containerID="6a2899a2bd48b07d11a54988ae27807805689a40bb41df1e769ca4da39298f4c" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.632459 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-s27fj"] Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.634196 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-s27fj" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.637499 4676 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-nt8rp" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.637900 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.638181 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.666051 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-s27fj"] Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.674340 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-jws6v"] Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.675039 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-jws6v" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.678663 4676 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-lzhxd" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.690216 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-jws6v"] Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.697711 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-w7kb7"] Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.698321 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.705881 4676 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-4p4cg" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.708311 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-w7kb7"] Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.755084 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckbvq\" (UniqueName: \"kubernetes.io/projected/1b51f181-5a85-4c07-b259-f67d17bf1134-kube-api-access-ckbvq\") pod \"cert-manager-858654f9db-jws6v\" (UID: \"1b51f181-5a85-4c07-b259-f67d17bf1134\") " pod="cert-manager/cert-manager-858654f9db-jws6v" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.755163 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n5fk\" (UniqueName: \"kubernetes.io/projected/6db681a1-6165-4404-81a9-a189e9b30bfd-kube-api-access-5n5fk\") pod \"cert-manager-cainjector-cf98fcc89-s27fj\" (UID: \"6db681a1-6165-4404-81a9-a189e9b30bfd\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-s27fj" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.856560 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n5fk\" (UniqueName: \"kubernetes.io/projected/6db681a1-6165-4404-81a9-a189e9b30bfd-kube-api-access-5n5fk\") pod \"cert-manager-cainjector-cf98fcc89-s27fj\" (UID: \"6db681a1-6165-4404-81a9-a189e9b30bfd\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-s27fj" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.856689 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckbvq\" (UniqueName: \"kubernetes.io/projected/1b51f181-5a85-4c07-b259-f67d17bf1134-kube-api-access-ckbvq\") pod \"cert-manager-858654f9db-jws6v\" (UID: \"1b51f181-5a85-4c07-b259-f67d17bf1134\") " pod="cert-manager/cert-manager-858654f9db-jws6v" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.856826 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk42x\" (UniqueName: \"kubernetes.io/projected/8f43788b-d559-4623-ae87-a820a2f23b08-kube-api-access-qk42x\") pod \"cert-manager-webhook-687f57d79b-w7kb7\" (UID: \"8f43788b-d559-4623-ae87-a820a2f23b08\") " pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.878231 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckbvq\" (UniqueName: \"kubernetes.io/projected/1b51f181-5a85-4c07-b259-f67d17bf1134-kube-api-access-ckbvq\") pod \"cert-manager-858654f9db-jws6v\" (UID: \"1b51f181-5a85-4c07-b259-f67d17bf1134\") " pod="cert-manager/cert-manager-858654f9db-jws6v" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.880172 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n5fk\" (UniqueName: \"kubernetes.io/projected/6db681a1-6165-4404-81a9-a189e9b30bfd-kube-api-access-5n5fk\") pod \"cert-manager-cainjector-cf98fcc89-s27fj\" (UID: \"6db681a1-6165-4404-81a9-a189e9b30bfd\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-s27fj" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.958319 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk42x\" (UniqueName: \"kubernetes.io/projected/8f43788b-d559-4623-ae87-a820a2f23b08-kube-api-access-qk42x\") pod \"cert-manager-webhook-687f57d79b-w7kb7\" (UID: \"8f43788b-d559-4623-ae87-a820a2f23b08\") " pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.966169 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-s27fj" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.974712 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk42x\" (UniqueName: \"kubernetes.io/projected/8f43788b-d559-4623-ae87-a820a2f23b08-kube-api-access-qk42x\") pod \"cert-manager-webhook-687f57d79b-w7kb7\" (UID: \"8f43788b-d559-4623-ae87-a820a2f23b08\") " pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" Jan 24 00:14:02 crc kubenswrapper[4676]: I0124 00:14:02.997666 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-jws6v" Jan 24 00:14:03 crc kubenswrapper[4676]: I0124 00:14:03.013987 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" Jan 24 00:14:03 crc kubenswrapper[4676]: I0124 00:14:03.226547 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-s27fj"] Jan 24 00:14:03 crc kubenswrapper[4676]: I0124 00:14:03.240917 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:14:03 crc kubenswrapper[4676]: I0124 00:14:03.250226 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-jws6v"] Jan 24 00:14:03 crc kubenswrapper[4676]: W0124 00:14:03.297216 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f43788b_d559_4623_ae87_a820a2f23b08.slice/crio-a36ce2c976d9ee609443c3d2f4e362184a38e380dfc12f6bcfac87fa11a03815 WatchSource:0}: Error finding container a36ce2c976d9ee609443c3d2f4e362184a38e380dfc12f6bcfac87fa11a03815: Status 404 returned error can't find the container with id a36ce2c976d9ee609443c3d2f4e362184a38e380dfc12f6bcfac87fa11a03815 Jan 24 00:14:03 crc kubenswrapper[4676]: I0124 00:14:03.297472 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-w7kb7"] Jan 24 00:14:03 crc kubenswrapper[4676]: I0124 00:14:03.315876 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" event={"ID":"8f43788b-d559-4623-ae87-a820a2f23b08","Type":"ContainerStarted","Data":"a36ce2c976d9ee609443c3d2f4e362184a38e380dfc12f6bcfac87fa11a03815"} Jan 24 00:14:03 crc kubenswrapper[4676]: I0124 00:14:03.320726 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-s27fj" event={"ID":"6db681a1-6165-4404-81a9-a189e9b30bfd","Type":"ContainerStarted","Data":"dab150e50614a7fcfffb7fde04ee0c5eb4c3aa33cf6329ff198bf4b3cd33a5e0"} Jan 24 00:14:03 crc kubenswrapper[4676]: I0124 00:14:03.321802 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-jws6v" event={"ID":"1b51f181-5a85-4c07-b259-f67d17bf1134","Type":"ContainerStarted","Data":"6e36bcc528299e3c444b1a5b51b721248634e9711d864189eb9cafb21be8688d"} Jan 24 00:14:08 crc kubenswrapper[4676]: I0124 00:14:08.369072 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" event={"ID":"8f43788b-d559-4623-ae87-a820a2f23b08","Type":"ContainerStarted","Data":"01173b1b3bc11b38321967d4f8e929aa6578c9495d169dccbf72bc4d4657304d"} Jan 24 00:14:08 crc kubenswrapper[4676]: I0124 00:14:08.370760 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" Jan 24 00:14:08 crc kubenswrapper[4676]: I0124 00:14:08.372472 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-s27fj" event={"ID":"6db681a1-6165-4404-81a9-a189e9b30bfd","Type":"ContainerStarted","Data":"53d6adf12612b975323d2b343cc3fc52b20c87621e4243827c8eb1ead8c1cab5"} Jan 24 00:14:08 crc kubenswrapper[4676]: I0124 00:14:08.375947 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-jws6v" event={"ID":"1b51f181-5a85-4c07-b259-f67d17bf1134","Type":"ContainerStarted","Data":"cc992135d2b376d78c382c95906e57272a00be1e59b87b4c440d6a4a84d13153"} Jan 24 00:14:08 crc kubenswrapper[4676]: I0124 00:14:08.398675 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" podStartSLOduration=2.349557902 podStartE2EDuration="6.398631643s" podCreationTimestamp="2026-01-24 00:14:02 +0000 UTC" firstStartedPulling="2026-01-24 00:14:03.300043045 +0000 UTC m=+627.330014046" lastFinishedPulling="2026-01-24 00:14:07.349116796 +0000 UTC m=+631.379087787" observedRunningTime="2026-01-24 00:14:08.392835946 +0000 UTC m=+632.422806947" watchObservedRunningTime="2026-01-24 00:14:08.398631643 +0000 UTC m=+632.428602694" Jan 24 00:14:08 crc kubenswrapper[4676]: I0124 00:14:08.425335 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-s27fj" podStartSLOduration=3.864665225 podStartE2EDuration="6.425299935s" podCreationTimestamp="2026-01-24 00:14:02 +0000 UTC" firstStartedPulling="2026-01-24 00:14:03.240510361 +0000 UTC m=+627.270481362" lastFinishedPulling="2026-01-24 00:14:05.801145031 +0000 UTC m=+629.831116072" observedRunningTime="2026-01-24 00:14:08.424158691 +0000 UTC m=+632.454129692" watchObservedRunningTime="2026-01-24 00:14:08.425299935 +0000 UTC m=+632.455270976" Jan 24 00:14:08 crc kubenswrapper[4676]: I0124 00:14:08.469914 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-jws6v" podStartSLOduration=2.384312211 podStartE2EDuration="6.469877614s" podCreationTimestamp="2026-01-24 00:14:02 +0000 UTC" firstStartedPulling="2026-01-24 00:14:03.264361988 +0000 UTC m=+627.294332989" lastFinishedPulling="2026-01-24 00:14:07.349927391 +0000 UTC m=+631.379898392" observedRunningTime="2026-01-24 00:14:08.449996598 +0000 UTC m=+632.479967599" watchObservedRunningTime="2026-01-24 00:14:08.469877614 +0000 UTC m=+632.499848625" Jan 24 00:14:12 crc kubenswrapper[4676]: I0124 00:14:12.544152 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ld569"] Jan 24 00:14:12 crc kubenswrapper[4676]: I0124 00:14:12.545368 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovn-controller" containerID="cri-o://a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468" gracePeriod=30 Jan 24 00:14:12 crc kubenswrapper[4676]: I0124 00:14:12.545483 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="nbdb" containerID="cri-o://0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58" gracePeriod=30 Jan 24 00:14:12 crc kubenswrapper[4676]: I0124 00:14:12.545552 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="sbdb" containerID="cri-o://a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624" gracePeriod=30 Jan 24 00:14:12 crc kubenswrapper[4676]: I0124 00:14:12.545622 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7" gracePeriod=30 Jan 24 00:14:12 crc kubenswrapper[4676]: I0124 00:14:12.545683 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kube-rbac-proxy-node" containerID="cri-o://02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc" gracePeriod=30 Jan 24 00:14:12 crc kubenswrapper[4676]: I0124 00:14:12.545743 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovn-acl-logging" containerID="cri-o://3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd" gracePeriod=30 Jan 24 00:14:12 crc kubenswrapper[4676]: I0124 00:14:12.545599 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="northd" containerID="cri-o://d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb" gracePeriod=30 Jan 24 00:14:12 crc kubenswrapper[4676]: I0124 00:14:12.604818 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" containerID="cri-o://b097984a7c57abb09d2fc1362781982b54bf19474644ee1742c87013905c7faf" gracePeriod=30 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.017566 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-w7kb7" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.412846 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/2.log" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.414172 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/1.log" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.414220 4676 generic.go:334] "Generic (PLEG): container finished" podID="b88e9d2e-35da-45a8-ac7e-22afd660ff9f" containerID="503448e193566525ada0f32c12c8a2978a0f18fbc763208a99e7e6534727cec5" exitCode=2 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.414299 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x57xf" event={"ID":"b88e9d2e-35da-45a8-ac7e-22afd660ff9f","Type":"ContainerDied","Data":"503448e193566525ada0f32c12c8a2978a0f18fbc763208a99e7e6534727cec5"} Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.414424 4676 scope.go:117] "RemoveContainer" containerID="cf92889c765992ceabf09d2de008fbbbfc1dc097012d57ce03aafee751eb759b" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.414803 4676 scope.go:117] "RemoveContainer" containerID="503448e193566525ada0f32c12c8a2978a0f18fbc763208a99e7e6534727cec5" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.414997 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x57xf_openshift-multus(b88e9d2e-35da-45a8-ac7e-22afd660ff9f)\"" pod="openshift-multus/multus-x57xf" podUID="b88e9d2e-35da-45a8-ac7e-22afd660ff9f" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.419670 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/3.log" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.426604 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovn-acl-logging/0.log" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427234 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovn-controller/0.log" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427778 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="b097984a7c57abb09d2fc1362781982b54bf19474644ee1742c87013905c7faf" exitCode=0 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427817 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624" exitCode=0 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427835 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58" exitCode=0 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427852 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb" exitCode=0 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427851 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"b097984a7c57abb09d2fc1362781982b54bf19474644ee1742c87013905c7faf"} Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427909 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624"} Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427928 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58"} Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427945 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb"} Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427968 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7"} Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.427867 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7" exitCode=0 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.428002 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc" exitCode=0 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.428022 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd" exitCode=143 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.428033 4676 generic.go:334] "Generic (PLEG): container finished" podID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerID="a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468" exitCode=143 Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.428054 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc"} Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.428068 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd"} Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.428082 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468"} Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.439680 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovnkube-controller/3.log" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.444566 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovn-acl-logging/0.log" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.445904 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovn-controller/0.log" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.446548 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.485890 4676 scope.go:117] "RemoveContainer" containerID="dc41985bef5146e5e21b5354222ebacb6310fa940511f64524b296d99bbd73e9" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534266 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c2dqg"] Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534612 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovn-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534630 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovn-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534642 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534649 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534659 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovn-acl-logging" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534667 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovn-acl-logging" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534677 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534683 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534694 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534700 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534717 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kubecfg-setup" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534725 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kubecfg-setup" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534733 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="sbdb" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534741 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="sbdb" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534754 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kube-rbac-proxy-ovn-metrics" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534762 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kube-rbac-proxy-ovn-metrics" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534770 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="nbdb" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534776 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="nbdb" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534787 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kube-rbac-proxy-node" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534794 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kube-rbac-proxy-node" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.534802 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="northd" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534807 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="northd" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534955 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="sbdb" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534972 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="northd" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534988 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kube-rbac-proxy-ovn-metrics" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.534997 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535006 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535017 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535030 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovn-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535039 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="kube-rbac-proxy-node" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535048 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535058 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovn-acl-logging" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535068 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535080 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="nbdb" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.535214 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535224 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: E0124 00:14:13.535234 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.535240 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" containerName="ovnkube-controller" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.537421 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.624971 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625024 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-openvswitch\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625053 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-kubelet\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625089 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4frqp\" (UniqueName: \"kubernetes.io/projected/24f0dc26-0857-430f-aebd-073fcfcc1c0a-kube-api-access-4frqp\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625126 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-ovn\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625141 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-log-socket\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625177 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-script-lib\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625152 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625198 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-var-lib-openvswitch\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625215 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-slash\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625241 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-bin\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625212 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625274 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625285 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovn-node-metrics-cert\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625246 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625313 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-systemd\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625334 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-etc-openvswitch\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625344 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625364 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-netns\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625403 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-node-log\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625432 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-env-overrides\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625456 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-ovn-kubernetes\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625477 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-systemd-units\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625505 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-netd\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625538 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-config\") pod \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\" (UID: \"24f0dc26-0857-430f-aebd-073fcfcc1c0a\") " Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625675 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-run-systemd\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625714 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-slash\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625733 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-run-openvswitch\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625757 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-cni-netd\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625773 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpx9m\" (UniqueName: \"kubernetes.io/projected/b082801b-819b-4d2d-835b-38a93b997fce-kube-api-access-hpx9m\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625791 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-kubelet\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625810 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625829 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-systemd-units\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625851 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b082801b-819b-4d2d-835b-38a93b997fce-ovnkube-config\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625875 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-run-ovn\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625905 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-run-ovn-kubernetes\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625926 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-cni-bin\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625941 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b082801b-819b-4d2d-835b-38a93b997fce-env-overrides\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625959 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-etc-openvswitch\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625984 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b082801b-819b-4d2d-835b-38a93b997fce-ovn-node-metrics-cert\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626002 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-var-lib-openvswitch\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626029 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-node-log\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626051 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b082801b-819b-4d2d-835b-38a93b997fce-ovnkube-script-lib\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626073 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-run-netns\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626092 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-log-socket\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626129 4676 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626140 4676 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626152 4676 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626164 4676 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626176 4676 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.625402 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-log-socket" (OuterVolumeSpecName: "log-socket") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626137 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626191 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-slash" (OuterVolumeSpecName: "host-slash") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626214 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626653 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626669 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626698 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626718 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626808 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626844 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.626905 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-node-log" (OuterVolumeSpecName: "node-log") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.627087 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.632944 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f0dc26-0857-430f-aebd-073fcfcc1c0a-kube-api-access-4frqp" (OuterVolumeSpecName: "kube-api-access-4frqp") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "kube-api-access-4frqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.634163 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.641624 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "24f0dc26-0857-430f-aebd-073fcfcc1c0a" (UID: "24f0dc26-0857-430f-aebd-073fcfcc1c0a"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727318 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-cni-netd\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727474 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-cni-netd\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727493 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-run-openvswitch\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727556 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpx9m\" (UniqueName: \"kubernetes.io/projected/b082801b-819b-4d2d-835b-38a93b997fce-kube-api-access-hpx9m\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727579 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-kubelet\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727585 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-run-openvswitch\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727606 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727633 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-systemd-units\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727650 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-kubelet\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727670 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b082801b-819b-4d2d-835b-38a93b997fce-ovnkube-config\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727703 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727720 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-run-ovn\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727745 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-run-ovn-kubernetes\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727753 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-systemd-units\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727776 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-cni-bin\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727794 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-etc-openvswitch\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727800 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-run-ovn\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727814 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b082801b-819b-4d2d-835b-38a93b997fce-env-overrides\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727853 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b082801b-819b-4d2d-835b-38a93b997fce-ovn-node-metrics-cert\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727878 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-var-lib-openvswitch\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727947 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-node-log\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.727998 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b082801b-819b-4d2d-835b-38a93b997fce-ovnkube-script-lib\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728034 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-run-netns\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728062 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-log-socket\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728095 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-run-systemd\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728141 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-slash\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728199 4676 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728211 4676 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728221 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4frqp\" (UniqueName: \"kubernetes.io/projected/24f0dc26-0857-430f-aebd-073fcfcc1c0a-kube-api-access-4frqp\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728231 4676 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-log-socket\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728240 4676 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728249 4676 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-slash\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728256 4676 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728268 4676 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/24f0dc26-0857-430f-aebd-073fcfcc1c0a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728276 4676 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728285 4676 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728294 4676 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728302 4676 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-node-log\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728311 4676 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/24f0dc26-0857-430f-aebd-073fcfcc1c0a-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728320 4676 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728330 4676 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/24f0dc26-0857-430f-aebd-073fcfcc1c0a-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728356 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-slash\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728418 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-run-ovn-kubernetes\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728449 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-cni-bin\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728476 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-etc-openvswitch\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728592 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-var-lib-openvswitch\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728796 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-log-socket\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728876 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-host-run-netns\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728917 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-run-systemd\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.728934 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b082801b-819b-4d2d-835b-38a93b997fce-node-log\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.729021 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b082801b-819b-4d2d-835b-38a93b997fce-ovnkube-config\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.729098 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b082801b-819b-4d2d-835b-38a93b997fce-env-overrides\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.729418 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b082801b-819b-4d2d-835b-38a93b997fce-ovnkube-script-lib\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.734122 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b082801b-819b-4d2d-835b-38a93b997fce-ovn-node-metrics-cert\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.751724 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpx9m\" (UniqueName: \"kubernetes.io/projected/b082801b-819b-4d2d-835b-38a93b997fce-kube-api-access-hpx9m\") pod \"ovnkube-node-c2dqg\" (UID: \"b082801b-819b-4d2d-835b-38a93b997fce\") " pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:13 crc kubenswrapper[4676]: I0124 00:14:13.854359 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.439984 4676 generic.go:334] "Generic (PLEG): container finished" podID="b082801b-819b-4d2d-835b-38a93b997fce" containerID="4fed15ba15b6e85c16ab1b2784cfd203e0efc654b51c36133af8d2ef84d4becb" exitCode=0 Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.440414 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerDied","Data":"4fed15ba15b6e85c16ab1b2784cfd203e0efc654b51c36133af8d2ef84d4becb"} Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.440480 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerStarted","Data":"daa79d40639e60df9f72aa02df9f92b56c6e10ba2812c60726eaf037482092c7"} Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.446215 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/2.log" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.454354 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovn-acl-logging/0.log" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.455198 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-ld569_24f0dc26-0857-430f-aebd-073fcfcc1c0a/ovn-controller/0.log" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.456061 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" event={"ID":"24f0dc26-0857-430f-aebd-073fcfcc1c0a","Type":"ContainerDied","Data":"a4ecac14869f9e61b6cf328f1a06fe9b463abccd1352f43143e709699b48fbd7"} Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.456126 4676 scope.go:117] "RemoveContainer" containerID="b097984a7c57abb09d2fc1362781982b54bf19474644ee1742c87013905c7faf" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.456363 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ld569" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.485129 4676 scope.go:117] "RemoveContainer" containerID="a1ccf0f4689bd5b3f634716a02e2c504e9cf4a1ebf5d95d06e1726133f4b2624" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.523351 4676 scope.go:117] "RemoveContainer" containerID="0c11b0bf64a540088ea316e492b601758f34a1a11e78622c5c084804b7213c58" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.569880 4676 scope.go:117] "RemoveContainer" containerID="d5580f29ee3a76e3eb08133c85be9d6a05b2738b900cd45b31c4fff775dab9bb" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.580191 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ld569"] Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.584449 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ld569"] Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.590123 4676 scope.go:117] "RemoveContainer" containerID="97d78e2b53d638374e7271129c05aa5b21f56dbc2abd4213f314f5a9220ad3c7" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.608019 4676 scope.go:117] "RemoveContainer" containerID="02d161fdcbf3861e821058380babbbf7ebb6a5929199df6285c046bed8d4d9cc" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.625106 4676 scope.go:117] "RemoveContainer" containerID="3878ebc66d4bc240cb14e18ed0dd1d1a06e65b3340a42aa0a54d70b5225422dd" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.651978 4676 scope.go:117] "RemoveContainer" containerID="a3d45db1c4c5e94ab3d5c20fc015bf49cce8f4306d7a56bead500d7ea13bb468" Jan 24 00:14:14 crc kubenswrapper[4676]: I0124 00:14:14.671678 4676 scope.go:117] "RemoveContainer" containerID="c59614eb0966d467422d52077d3fcb569d7c66e7b1ce142a7c2b3a548c315551" Jan 24 00:14:15 crc kubenswrapper[4676]: I0124 00:14:15.466733 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerStarted","Data":"4bbb3007fa5db6b08b3a2e059bbdcabd98a314f78f1b8bd2145c9c8679a3c131"} Jan 24 00:14:15 crc kubenswrapper[4676]: I0124 00:14:15.467218 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerStarted","Data":"0019b481bcecfa427cbcbd8601b9929cba76fc95252a31658dc859a343bf16f1"} Jan 24 00:14:15 crc kubenswrapper[4676]: I0124 00:14:15.467233 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerStarted","Data":"31df2e60c29211f17eb1cb0a277d02f598cb618a745a4067410dd3d8a32b0283"} Jan 24 00:14:15 crc kubenswrapper[4676]: I0124 00:14:15.467244 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerStarted","Data":"e8c2615fd559f2b8bbe3f170b11362efef09bf752021d8c0c06f5930919dcea9"} Jan 24 00:14:15 crc kubenswrapper[4676]: I0124 00:14:15.467305 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerStarted","Data":"57ca41b08691a5e2d17bac2255d196d1a86559a93b43a71f6fc2d04a24f8fade"} Jan 24 00:14:15 crc kubenswrapper[4676]: I0124 00:14:15.467316 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerStarted","Data":"6088a3bd55a04f084e276574e48ff5ad3268ac83bd6ff1b2e3eead89794d8df0"} Jan 24 00:14:16 crc kubenswrapper[4676]: I0124 00:14:16.263533 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24f0dc26-0857-430f-aebd-073fcfcc1c0a" path="/var/lib/kubelet/pods/24f0dc26-0857-430f-aebd-073fcfcc1c0a/volumes" Jan 24 00:14:18 crc kubenswrapper[4676]: I0124 00:14:18.494519 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerStarted","Data":"92fd32939e1ff79efddb46e5b0b57cbd61f5e788a9337e2c5b375b985959b6a3"} Jan 24 00:14:20 crc kubenswrapper[4676]: I0124 00:14:20.511198 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" event={"ID":"b082801b-819b-4d2d-835b-38a93b997fce","Type":"ContainerStarted","Data":"d408ca0bb7091ab554b81d1909a3d5528818046f1e5197886616b8b1040f98f8"} Jan 24 00:14:20 crc kubenswrapper[4676]: I0124 00:14:20.511812 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:20 crc kubenswrapper[4676]: I0124 00:14:20.511946 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:20 crc kubenswrapper[4676]: I0124 00:14:20.555916 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:20 crc kubenswrapper[4676]: I0124 00:14:20.613128 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" podStartSLOduration=7.613110871 podStartE2EDuration="7.613110871s" podCreationTimestamp="2026-01-24 00:14:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:14:20.569306407 +0000 UTC m=+644.599277418" watchObservedRunningTime="2026-01-24 00:14:20.613110871 +0000 UTC m=+644.643081882" Jan 24 00:14:21 crc kubenswrapper[4676]: I0124 00:14:21.520703 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:21 crc kubenswrapper[4676]: I0124 00:14:21.566507 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:26 crc kubenswrapper[4676]: I0124 00:14:26.260538 4676 scope.go:117] "RemoveContainer" containerID="503448e193566525ada0f32c12c8a2978a0f18fbc763208a99e7e6534727cec5" Jan 24 00:14:26 crc kubenswrapper[4676]: E0124 00:14:26.261665 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-x57xf_openshift-multus(b88e9d2e-35da-45a8-ac7e-22afd660ff9f)\"" pod="openshift-multus/multus-x57xf" podUID="b88e9d2e-35da-45a8-ac7e-22afd660ff9f" Jan 24 00:14:39 crc kubenswrapper[4676]: I0124 00:14:39.255642 4676 scope.go:117] "RemoveContainer" containerID="503448e193566525ada0f32c12c8a2978a0f18fbc763208a99e7e6534727cec5" Jan 24 00:14:39 crc kubenswrapper[4676]: I0124 00:14:39.657742 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-x57xf_b88e9d2e-35da-45a8-ac7e-22afd660ff9f/kube-multus/2.log" Jan 24 00:14:39 crc kubenswrapper[4676]: I0124 00:14:39.657809 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-x57xf" event={"ID":"b88e9d2e-35da-45a8-ac7e-22afd660ff9f","Type":"ContainerStarted","Data":"088ce398c4ed98e98c167dcb59af037593455aada9daa8c18b5830c108862309"} Jan 24 00:14:43 crc kubenswrapper[4676]: I0124 00:14:43.887506 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c2dqg" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.466309 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv"] Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.467842 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.469464 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.477857 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv"] Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.600152 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.600196 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.600257 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdrc2\" (UniqueName: \"kubernetes.io/projected/79811b63-a3e6-47b0-8041-247b1536ba50-kube-api-access-vdrc2\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.701415 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.701810 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.702042 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.702288 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdrc2\" (UniqueName: \"kubernetes.io/projected/79811b63-a3e6-47b0-8041-247b1536ba50-kube-api-access-vdrc2\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.702544 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.728085 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdrc2\" (UniqueName: \"kubernetes.io/projected/79811b63-a3e6-47b0-8041-247b1536ba50-kube-api-access-vdrc2\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:54 crc kubenswrapper[4676]: I0124 00:14:54.783952 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:14:55 crc kubenswrapper[4676]: I0124 00:14:55.023158 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv"] Jan 24 00:14:55 crc kubenswrapper[4676]: I0124 00:14:55.763668 4676 generic.go:334] "Generic (PLEG): container finished" podID="79811b63-a3e6-47b0-8041-247b1536ba50" containerID="854c4496f04919c2bab64d3d7604ac339ddfc6ba556e38b6d39a05208f375eb4" exitCode=0 Jan 24 00:14:55 crc kubenswrapper[4676]: I0124 00:14:55.763754 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" event={"ID":"79811b63-a3e6-47b0-8041-247b1536ba50","Type":"ContainerDied","Data":"854c4496f04919c2bab64d3d7604ac339ddfc6ba556e38b6d39a05208f375eb4"} Jan 24 00:14:55 crc kubenswrapper[4676]: I0124 00:14:55.764208 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" event={"ID":"79811b63-a3e6-47b0-8041-247b1536ba50","Type":"ContainerStarted","Data":"2f8fc73d1311811b964eaef9ece132796c7141b06fe05cb4addc6171b93ccaed"} Jan 24 00:14:57 crc kubenswrapper[4676]: I0124 00:14:57.778263 4676 generic.go:334] "Generic (PLEG): container finished" podID="79811b63-a3e6-47b0-8041-247b1536ba50" containerID="efe746eccedd511dacaaf4516927d95056452db8cbdcc622f804b5020b8c32ea" exitCode=0 Jan 24 00:14:57 crc kubenswrapper[4676]: I0124 00:14:57.778404 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" event={"ID":"79811b63-a3e6-47b0-8041-247b1536ba50","Type":"ContainerDied","Data":"efe746eccedd511dacaaf4516927d95056452db8cbdcc622f804b5020b8c32ea"} Jan 24 00:14:58 crc kubenswrapper[4676]: I0124 00:14:58.792130 4676 generic.go:334] "Generic (PLEG): container finished" podID="79811b63-a3e6-47b0-8041-247b1536ba50" containerID="8477a53dd8b97bb84ef79ddf7cc0b2f8265d75e677d413a8a09a26e547220989" exitCode=0 Jan 24 00:14:58 crc kubenswrapper[4676]: I0124 00:14:58.792546 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" event={"ID":"79811b63-a3e6-47b0-8041-247b1536ba50","Type":"ContainerDied","Data":"8477a53dd8b97bb84ef79ddf7cc0b2f8265d75e677d413a8a09a26e547220989"} Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.101972 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.179470 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p"] Jan 24 00:15:00 crc kubenswrapper[4676]: E0124 00:15:00.179734 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79811b63-a3e6-47b0-8041-247b1536ba50" containerName="util" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.179754 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="79811b63-a3e6-47b0-8041-247b1536ba50" containerName="util" Jan 24 00:15:00 crc kubenswrapper[4676]: E0124 00:15:00.179770 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79811b63-a3e6-47b0-8041-247b1536ba50" containerName="pull" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.179778 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="79811b63-a3e6-47b0-8041-247b1536ba50" containerName="pull" Jan 24 00:15:00 crc kubenswrapper[4676]: E0124 00:15:00.179799 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79811b63-a3e6-47b0-8041-247b1536ba50" containerName="extract" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.179807 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="79811b63-a3e6-47b0-8041-247b1536ba50" containerName="extract" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.179931 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="79811b63-a3e6-47b0-8041-247b1536ba50" containerName="extract" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.180464 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.184135 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.184612 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.187171 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p"] Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.187618 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-bundle\") pod \"79811b63-a3e6-47b0-8041-247b1536ba50\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.187922 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-util\") pod \"79811b63-a3e6-47b0-8041-247b1536ba50\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.187987 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdrc2\" (UniqueName: \"kubernetes.io/projected/79811b63-a3e6-47b0-8041-247b1536ba50-kube-api-access-vdrc2\") pod \"79811b63-a3e6-47b0-8041-247b1536ba50\" (UID: \"79811b63-a3e6-47b0-8041-247b1536ba50\") " Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.189077 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-bundle" (OuterVolumeSpecName: "bundle") pod "79811b63-a3e6-47b0-8041-247b1536ba50" (UID: "79811b63-a3e6-47b0-8041-247b1536ba50"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.204641 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79811b63-a3e6-47b0-8041-247b1536ba50-kube-api-access-vdrc2" (OuterVolumeSpecName: "kube-api-access-vdrc2") pod "79811b63-a3e6-47b0-8041-247b1536ba50" (UID: "79811b63-a3e6-47b0-8041-247b1536ba50"). InnerVolumeSpecName "kube-api-access-vdrc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.212561 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-util" (OuterVolumeSpecName: "util") pod "79811b63-a3e6-47b0-8041-247b1536ba50" (UID: "79811b63-a3e6-47b0-8041-247b1536ba50"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.290067 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ef8e58-50e2-4cc5-b92c-35710605e5b1-secret-volume\") pod \"collect-profiles-29486895-5v67p\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.290140 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bvwt\" (UniqueName: \"kubernetes.io/projected/40ef8e58-50e2-4cc5-b92c-35710605e5b1-kube-api-access-7bvwt\") pod \"collect-profiles-29486895-5v67p\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.290185 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ef8e58-50e2-4cc5-b92c-35710605e5b1-config-volume\") pod \"collect-profiles-29486895-5v67p\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.290280 4676 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-util\") on node \"crc\" DevicePath \"\"" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.290298 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdrc2\" (UniqueName: \"kubernetes.io/projected/79811b63-a3e6-47b0-8041-247b1536ba50-kube-api-access-vdrc2\") on node \"crc\" DevicePath \"\"" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.290314 4676 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/79811b63-a3e6-47b0-8041-247b1536ba50-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.391005 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ef8e58-50e2-4cc5-b92c-35710605e5b1-config-volume\") pod \"collect-profiles-29486895-5v67p\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.391101 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ef8e58-50e2-4cc5-b92c-35710605e5b1-secret-volume\") pod \"collect-profiles-29486895-5v67p\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.391185 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bvwt\" (UniqueName: \"kubernetes.io/projected/40ef8e58-50e2-4cc5-b92c-35710605e5b1-kube-api-access-7bvwt\") pod \"collect-profiles-29486895-5v67p\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.392216 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ef8e58-50e2-4cc5-b92c-35710605e5b1-config-volume\") pod \"collect-profiles-29486895-5v67p\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.397389 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ef8e58-50e2-4cc5-b92c-35710605e5b1-secret-volume\") pod \"collect-profiles-29486895-5v67p\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.417810 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bvwt\" (UniqueName: \"kubernetes.io/projected/40ef8e58-50e2-4cc5-b92c-35710605e5b1-kube-api-access-7bvwt\") pod \"collect-profiles-29486895-5v67p\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.494605 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.751504 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p"] Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.814151 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" event={"ID":"79811b63-a3e6-47b0-8041-247b1536ba50","Type":"ContainerDied","Data":"2f8fc73d1311811b964eaef9ece132796c7141b06fe05cb4addc6171b93ccaed"} Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.814503 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f8fc73d1311811b964eaef9ece132796c7141b06fe05cb4addc6171b93ccaed" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.814627 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv" Jan 24 00:15:00 crc kubenswrapper[4676]: I0124 00:15:00.818020 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" event={"ID":"40ef8e58-50e2-4cc5-b92c-35710605e5b1","Type":"ContainerStarted","Data":"4e0a09189da28f9a231ef5cf2463bb087d76d53a5efd543b3017b1dc4cbb92ba"} Jan 24 00:15:01 crc kubenswrapper[4676]: I0124 00:15:01.826336 4676 generic.go:334] "Generic (PLEG): container finished" podID="40ef8e58-50e2-4cc5-b92c-35710605e5b1" containerID="1f9529fca33e87b8d8a2cf842fe194d301e46a27673eb8a957977d0bcb8f4f8d" exitCode=0 Jan 24 00:15:01 crc kubenswrapper[4676]: I0124 00:15:01.826412 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" event={"ID":"40ef8e58-50e2-4cc5-b92c-35710605e5b1","Type":"ContainerDied","Data":"1f9529fca33e87b8d8a2cf842fe194d301e46a27673eb8a957977d0bcb8f4f8d"} Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.016861 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.107422 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-jsbxx"] Jan 24 00:15:03 crc kubenswrapper[4676]: E0124 00:15:03.107660 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40ef8e58-50e2-4cc5-b92c-35710605e5b1" containerName="collect-profiles" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.107671 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="40ef8e58-50e2-4cc5-b92c-35710605e5b1" containerName="collect-profiles" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.107795 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="40ef8e58-50e2-4cc5-b92c-35710605e5b1" containerName="collect-profiles" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.108204 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-jsbxx" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.110575 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-fpvgz" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.110612 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.110684 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.123410 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-jsbxx"] Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.130537 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ef8e58-50e2-4cc5-b92c-35710605e5b1-secret-volume\") pod \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.130632 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ef8e58-50e2-4cc5-b92c-35710605e5b1-config-volume\") pod \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.130670 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bvwt\" (UniqueName: \"kubernetes.io/projected/40ef8e58-50e2-4cc5-b92c-35710605e5b1-kube-api-access-7bvwt\") pod \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\" (UID: \"40ef8e58-50e2-4cc5-b92c-35710605e5b1\") " Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.131414 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40ef8e58-50e2-4cc5-b92c-35710605e5b1-config-volume" (OuterVolumeSpecName: "config-volume") pod "40ef8e58-50e2-4cc5-b92c-35710605e5b1" (UID: "40ef8e58-50e2-4cc5-b92c-35710605e5b1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.136898 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40ef8e58-50e2-4cc5-b92c-35710605e5b1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "40ef8e58-50e2-4cc5-b92c-35710605e5b1" (UID: "40ef8e58-50e2-4cc5-b92c-35710605e5b1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.154602 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40ef8e58-50e2-4cc5-b92c-35710605e5b1-kube-api-access-7bvwt" (OuterVolumeSpecName: "kube-api-access-7bvwt") pod "40ef8e58-50e2-4cc5-b92c-35710605e5b1" (UID: "40ef8e58-50e2-4cc5-b92c-35710605e5b1"). InnerVolumeSpecName "kube-api-access-7bvwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.233084 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9qjn\" (UniqueName: \"kubernetes.io/projected/efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2-kube-api-access-t9qjn\") pod \"nmstate-operator-646758c888-jsbxx\" (UID: \"efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2\") " pod="openshift-nmstate/nmstate-operator-646758c888-jsbxx" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.233227 4676 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/40ef8e58-50e2-4cc5-b92c-35710605e5b1-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.233247 4676 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/40ef8e58-50e2-4cc5-b92c-35710605e5b1-config-volume\") on node \"crc\" DevicePath \"\"" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.233259 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bvwt\" (UniqueName: \"kubernetes.io/projected/40ef8e58-50e2-4cc5-b92c-35710605e5b1-kube-api-access-7bvwt\") on node \"crc\" DevicePath \"\"" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.334868 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9qjn\" (UniqueName: \"kubernetes.io/projected/efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2-kube-api-access-t9qjn\") pod \"nmstate-operator-646758c888-jsbxx\" (UID: \"efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2\") " pod="openshift-nmstate/nmstate-operator-646758c888-jsbxx" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.356435 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9qjn\" (UniqueName: \"kubernetes.io/projected/efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2-kube-api-access-t9qjn\") pod \"nmstate-operator-646758c888-jsbxx\" (UID: \"efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2\") " pod="openshift-nmstate/nmstate-operator-646758c888-jsbxx" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.419915 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-jsbxx" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.648882 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-jsbxx"] Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.837617 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-jsbxx" event={"ID":"efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2","Type":"ContainerStarted","Data":"92cccb2822eb5ab7dff341249774a2f78675e8dd3c31865ae18bc7415a408904"} Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.839138 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" event={"ID":"40ef8e58-50e2-4cc5-b92c-35710605e5b1","Type":"ContainerDied","Data":"4e0a09189da28f9a231ef5cf2463bb087d76d53a5efd543b3017b1dc4cbb92ba"} Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.839171 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e0a09189da28f9a231ef5cf2463bb087d76d53a5efd543b3017b1dc4cbb92ba" Jan 24 00:15:03 crc kubenswrapper[4676]: I0124 00:15:03.839228 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p" Jan 24 00:15:06 crc kubenswrapper[4676]: I0124 00:15:06.863612 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-jsbxx" event={"ID":"efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2","Type":"ContainerStarted","Data":"c6d3d119aa50a84f358aa734505a8382c68b2998b0fa588a9d3004b9bddda533"} Jan 24 00:15:06 crc kubenswrapper[4676]: I0124 00:15:06.890688 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-jsbxx" podStartSLOduration=1.595507754 podStartE2EDuration="3.890667203s" podCreationTimestamp="2026-01-24 00:15:03 +0000 UTC" firstStartedPulling="2026-01-24 00:15:03.66894515 +0000 UTC m=+687.698916151" lastFinishedPulling="2026-01-24 00:15:05.964104589 +0000 UTC m=+689.994075600" observedRunningTime="2026-01-24 00:15:06.886205706 +0000 UTC m=+690.916176747" watchObservedRunningTime="2026-01-24 00:15:06.890667203 +0000 UTC m=+690.920638214" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.348757 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-rv2q7"] Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.350464 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-rv2q7" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.353287 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv"] Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.353978 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.392080 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv"] Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.396366 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.398002 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-fgnjq" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.401005 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-rv2q7"] Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.413475 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-lkqsz"] Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.414589 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.511439 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-dbus-socket\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.511480 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-nmstate-lock\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.511504 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/badb3470-b60b-44ca-8d9e-52191ea016fa-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-cpbrv\" (UID: \"badb3470-b60b-44ca-8d9e-52191ea016fa\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.511529 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj6g7\" (UniqueName: \"kubernetes.io/projected/badb3470-b60b-44ca-8d9e-52191ea016fa-kube-api-access-fj6g7\") pod \"nmstate-webhook-8474b5b9d8-cpbrv\" (UID: \"badb3470-b60b-44ca-8d9e-52191ea016fa\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.511606 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdn9z\" (UniqueName: \"kubernetes.io/projected/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-kube-api-access-gdn9z\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.511622 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd422\" (UniqueName: \"kubernetes.io/projected/0c99caac-d9eb-494c-bd04-c18dbc8a0844-kube-api-access-dd422\") pod \"nmstate-metrics-54757c584b-rv2q7\" (UID: \"0c99caac-d9eb-494c-bd04-c18dbc8a0844\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-rv2q7" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.511638 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-ovs-socket\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.539462 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988"] Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.540086 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.541892 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.542546 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.547232 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-4znsj" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.548720 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988"] Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613463 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdn9z\" (UniqueName: \"kubernetes.io/projected/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-kube-api-access-gdn9z\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613500 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd422\" (UniqueName: \"kubernetes.io/projected/0c99caac-d9eb-494c-bd04-c18dbc8a0844-kube-api-access-dd422\") pod \"nmstate-metrics-54757c584b-rv2q7\" (UID: \"0c99caac-d9eb-494c-bd04-c18dbc8a0844\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-rv2q7" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613526 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-ovs-socket\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613556 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-dbus-socket\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613572 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-nmstate-lock\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613589 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/badb3470-b60b-44ca-8d9e-52191ea016fa-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-cpbrv\" (UID: \"badb3470-b60b-44ca-8d9e-52191ea016fa\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613614 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj6g7\" (UniqueName: \"kubernetes.io/projected/badb3470-b60b-44ca-8d9e-52191ea016fa-kube-api-access-fj6g7\") pod \"nmstate-webhook-8474b5b9d8-cpbrv\" (UID: \"badb3470-b60b-44ca-8d9e-52191ea016fa\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613674 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-ovs-socket\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613697 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-nmstate-lock\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.613913 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-dbus-socket\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.628672 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/badb3470-b60b-44ca-8d9e-52191ea016fa-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-cpbrv\" (UID: \"badb3470-b60b-44ca-8d9e-52191ea016fa\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.630362 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdn9z\" (UniqueName: \"kubernetes.io/projected/6bdbc24e-ee52-41f0-9aaf-1091ac803c27-kube-api-access-gdn9z\") pod \"nmstate-handler-lkqsz\" (UID: \"6bdbc24e-ee52-41f0-9aaf-1091ac803c27\") " pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.631692 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd422\" (UniqueName: \"kubernetes.io/projected/0c99caac-d9eb-494c-bd04-c18dbc8a0844-kube-api-access-dd422\") pod \"nmstate-metrics-54757c584b-rv2q7\" (UID: \"0c99caac-d9eb-494c-bd04-c18dbc8a0844\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-rv2q7" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.636082 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj6g7\" (UniqueName: \"kubernetes.io/projected/badb3470-b60b-44ca-8d9e-52191ea016fa-kube-api-access-fj6g7\") pod \"nmstate-webhook-8474b5b9d8-cpbrv\" (UID: \"badb3470-b60b-44ca-8d9e-52191ea016fa\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.699775 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-rv2q7" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.708985 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.718026 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4db34b09-85e6-435d-b991-c2513eec5d17-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-95988\" (UID: \"4db34b09-85e6-435d-b991-c2513eec5d17\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.718233 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4db34b09-85e6-435d-b991-c2513eec5d17-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-95988\" (UID: \"4db34b09-85e6-435d-b991-c2513eec5d17\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.718254 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpqm6\" (UniqueName: \"kubernetes.io/projected/4db34b09-85e6-435d-b991-c2513eec5d17-kube-api-access-lpqm6\") pod \"nmstate-console-plugin-7754f76f8b-95988\" (UID: \"4db34b09-85e6-435d-b991-c2513eec5d17\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.739246 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-646b8b7948-5mzrb"] Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.739889 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.750746 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.766430 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-646b8b7948-5mzrb"] Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.822032 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4db34b09-85e6-435d-b991-c2513eec5d17-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-95988\" (UID: \"4db34b09-85e6-435d-b991-c2513eec5d17\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.822068 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4db34b09-85e6-435d-b991-c2513eec5d17-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-95988\" (UID: \"4db34b09-85e6-435d-b991-c2513eec5d17\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.822088 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpqm6\" (UniqueName: \"kubernetes.io/projected/4db34b09-85e6-435d-b991-c2513eec5d17-kube-api-access-lpqm6\") pod \"nmstate-console-plugin-7754f76f8b-95988\" (UID: \"4db34b09-85e6-435d-b991-c2513eec5d17\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.823344 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/4db34b09-85e6-435d-b991-c2513eec5d17-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-95988\" (UID: \"4db34b09-85e6-435d-b991-c2513eec5d17\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.832122 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/4db34b09-85e6-435d-b991-c2513eec5d17-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-95988\" (UID: \"4db34b09-85e6-435d-b991-c2513eec5d17\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.853394 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpqm6\" (UniqueName: \"kubernetes.io/projected/4db34b09-85e6-435d-b991-c2513eec5d17-kube-api-access-lpqm6\") pod \"nmstate-console-plugin-7754f76f8b-95988\" (UID: \"4db34b09-85e6-435d-b991-c2513eec5d17\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.855871 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.923309 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-console-serving-cert\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.923349 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-console-config\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.923487 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgnzk\" (UniqueName: \"kubernetes.io/projected/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-kube-api-access-vgnzk\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.923527 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-trusted-ca-bundle\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.923618 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-service-ca\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.923645 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-console-oauth-config\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.923694 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-oauth-serving-cert\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:29 crc kubenswrapper[4676]: I0124 00:15:29.988226 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-rv2q7"] Jan 24 00:15:30 crc kubenswrapper[4676]: W0124 00:15:30.005756 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c99caac_d9eb_494c_bd04_c18dbc8a0844.slice/crio-5c90ef64ad9fae698aa4621f61cb972076b24d792ea61941543af01d0c927a6f WatchSource:0}: Error finding container 5c90ef64ad9fae698aa4621f61cb972076b24d792ea61941543af01d0c927a6f: Status 404 returned error can't find the container with id 5c90ef64ad9fae698aa4621f61cb972076b24d792ea61941543af01d0c927a6f Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.010287 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv"] Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.014826 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lkqsz" event={"ID":"6bdbc24e-ee52-41f0-9aaf-1091ac803c27","Type":"ContainerStarted","Data":"35175506727db5307573d9cc30023df1695d2911019380bc72f5197925af315f"} Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.015954 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-rv2q7" event={"ID":"0c99caac-d9eb-494c-bd04-c18dbc8a0844","Type":"ContainerStarted","Data":"5c90ef64ad9fae698aa4621f61cb972076b24d792ea61941543af01d0c927a6f"} Jan 24 00:15:30 crc kubenswrapper[4676]: W0124 00:15:30.017342 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbadb3470_b60b_44ca_8d9e_52191ea016fa.slice/crio-7bdf627a13b1ccbdffa4ec8f369a655cf8fb07762595b6dc01fa8326be4a9925 WatchSource:0}: Error finding container 7bdf627a13b1ccbdffa4ec8f369a655cf8fb07762595b6dc01fa8326be4a9925: Status 404 returned error can't find the container with id 7bdf627a13b1ccbdffa4ec8f369a655cf8fb07762595b6dc01fa8326be4a9925 Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.025672 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-console-oauth-config\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.026664 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-oauth-serving-cert\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.026706 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-console-config\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.026756 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-console-serving-cert\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.026845 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgnzk\" (UniqueName: \"kubernetes.io/projected/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-kube-api-access-vgnzk\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.026945 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-trusted-ca-bundle\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.026993 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-service-ca\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.027545 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-oauth-serving-cert\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.027778 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-service-ca\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.028343 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-console-config\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.028949 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-trusted-ca-bundle\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.029800 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-console-oauth-config\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.034674 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-console-serving-cert\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.043805 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgnzk\" (UniqueName: \"kubernetes.io/projected/19d2088d-e8c0-4c76-be55-b5a9d3d1e681-kube-api-access-vgnzk\") pod \"console-646b8b7948-5mzrb\" (UID: \"19d2088d-e8c0-4c76-be55-b5a9d3d1e681\") " pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.072964 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.077107 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988"] Jan 24 00:15:30 crc kubenswrapper[4676]: I0124 00:15:30.238668 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-646b8b7948-5mzrb"] Jan 24 00:15:30 crc kubenswrapper[4676]: W0124 00:15:30.241572 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19d2088d_e8c0_4c76_be55_b5a9d3d1e681.slice/crio-e30dee6d031278dadfaa4547c9699eb6b480f4a1dc38a0850888cb9f5a2fb383 WatchSource:0}: Error finding container e30dee6d031278dadfaa4547c9699eb6b480f4a1dc38a0850888cb9f5a2fb383: Status 404 returned error can't find the container with id e30dee6d031278dadfaa4547c9699eb6b480f4a1dc38a0850888cb9f5a2fb383 Jan 24 00:15:31 crc kubenswrapper[4676]: I0124 00:15:31.047875 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-646b8b7948-5mzrb" event={"ID":"19d2088d-e8c0-4c76-be55-b5a9d3d1e681","Type":"ContainerStarted","Data":"fba11e043b13fe9ba986b50fe5a6a225e513e9c6b83f1d17947e2dcfb50b0d38"} Jan 24 00:15:31 crc kubenswrapper[4676]: I0124 00:15:31.048413 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-646b8b7948-5mzrb" event={"ID":"19d2088d-e8c0-4c76-be55-b5a9d3d1e681","Type":"ContainerStarted","Data":"e30dee6d031278dadfaa4547c9699eb6b480f4a1dc38a0850888cb9f5a2fb383"} Jan 24 00:15:31 crc kubenswrapper[4676]: I0124 00:15:31.053502 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" event={"ID":"4db34b09-85e6-435d-b991-c2513eec5d17","Type":"ContainerStarted","Data":"0f0fa59f140e0b22263007c638a8d7342517e2280615defc8f3853c0688da093"} Jan 24 00:15:31 crc kubenswrapper[4676]: I0124 00:15:31.060700 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" event={"ID":"badb3470-b60b-44ca-8d9e-52191ea016fa","Type":"ContainerStarted","Data":"7bdf627a13b1ccbdffa4ec8f369a655cf8fb07762595b6dc01fa8326be4a9925"} Jan 24 00:15:31 crc kubenswrapper[4676]: I0124 00:15:31.086192 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-646b8b7948-5mzrb" podStartSLOduration=2.086135577 podStartE2EDuration="2.086135577s" podCreationTimestamp="2026-01-24 00:15:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:15:31.081987651 +0000 UTC m=+715.111958692" watchObservedRunningTime="2026-01-24 00:15:31.086135577 +0000 UTC m=+715.116106588" Jan 24 00:15:34 crc kubenswrapper[4676]: I0124 00:15:34.079320 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lkqsz" event={"ID":"6bdbc24e-ee52-41f0-9aaf-1091ac803c27","Type":"ContainerStarted","Data":"0aca8597998e206c4a952a56598aab99cefe88f62b281da6908483d638961f79"} Jan 24 00:15:34 crc kubenswrapper[4676]: I0124 00:15:34.079934 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:34 crc kubenswrapper[4676]: I0124 00:15:34.083415 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" event={"ID":"4db34b09-85e6-435d-b991-c2513eec5d17","Type":"ContainerStarted","Data":"a0e665983abe577c8bc2a4fbbaf9624fceff77ef0f6b40eacd5d5cd4f0eaf012"} Jan 24 00:15:34 crc kubenswrapper[4676]: I0124 00:15:34.085956 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" event={"ID":"badb3470-b60b-44ca-8d9e-52191ea016fa","Type":"ContainerStarted","Data":"9ac7a9454ee76866513183093ce28706ab20aa234cf23438a87dbbe2e43cbf9d"} Jan 24 00:15:34 crc kubenswrapper[4676]: I0124 00:15:34.086592 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:15:34 crc kubenswrapper[4676]: I0124 00:15:34.088824 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-rv2q7" event={"ID":"0c99caac-d9eb-494c-bd04-c18dbc8a0844","Type":"ContainerStarted","Data":"40d4597c38bafb120668cbf6b639ed71e2620fa0bfa8625c6f6e949c4aedf8c6"} Jan 24 00:15:34 crc kubenswrapper[4676]: I0124 00:15:34.102827 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-lkqsz" podStartSLOduration=1.922074602 podStartE2EDuration="5.102806941s" podCreationTimestamp="2026-01-24 00:15:29 +0000 UTC" firstStartedPulling="2026-01-24 00:15:29.799973474 +0000 UTC m=+713.829944475" lastFinishedPulling="2026-01-24 00:15:32.980705813 +0000 UTC m=+717.010676814" observedRunningTime="2026-01-24 00:15:34.100747727 +0000 UTC m=+718.130718748" watchObservedRunningTime="2026-01-24 00:15:34.102806941 +0000 UTC m=+718.132777952" Jan 24 00:15:34 crc kubenswrapper[4676]: I0124 00:15:34.167631 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" podStartSLOduration=2.209733919 podStartE2EDuration="5.167611032s" podCreationTimestamp="2026-01-24 00:15:29 +0000 UTC" firstStartedPulling="2026-01-24 00:15:30.019140668 +0000 UTC m=+714.049111669" lastFinishedPulling="2026-01-24 00:15:32.977017771 +0000 UTC m=+717.006988782" observedRunningTime="2026-01-24 00:15:34.121825472 +0000 UTC m=+718.151796493" watchObservedRunningTime="2026-01-24 00:15:34.167611032 +0000 UTC m=+718.197582033" Jan 24 00:15:34 crc kubenswrapper[4676]: I0124 00:15:34.169676 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-95988" podStartSLOduration=2.287322754 podStartE2EDuration="5.169666596s" podCreationTimestamp="2026-01-24 00:15:29 +0000 UTC" firstStartedPulling="2026-01-24 00:15:30.087260932 +0000 UTC m=+714.117231933" lastFinishedPulling="2026-01-24 00:15:32.969604754 +0000 UTC m=+716.999575775" observedRunningTime="2026-01-24 00:15:34.165990623 +0000 UTC m=+718.195961634" watchObservedRunningTime="2026-01-24 00:15:34.169666596 +0000 UTC m=+718.199637597" Jan 24 00:15:36 crc kubenswrapper[4676]: I0124 00:15:36.103190 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-rv2q7" event={"ID":"0c99caac-d9eb-494c-bd04-c18dbc8a0844","Type":"ContainerStarted","Data":"22e7763a5ed5b606f7dc1ac423a90590967569c2253c1453d63ea52f3965c491"} Jan 24 00:15:36 crc kubenswrapper[4676]: I0124 00:15:36.129775 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-rv2q7" podStartSLOduration=1.595215281 podStartE2EDuration="7.129751627s" podCreationTimestamp="2026-01-24 00:15:29 +0000 UTC" firstStartedPulling="2026-01-24 00:15:30.009921796 +0000 UTC m=+714.039892797" lastFinishedPulling="2026-01-24 00:15:35.544458142 +0000 UTC m=+719.574429143" observedRunningTime="2026-01-24 00:15:36.125966771 +0000 UTC m=+720.155937812" watchObservedRunningTime="2026-01-24 00:15:36.129751627 +0000 UTC m=+720.159722668" Jan 24 00:15:39 crc kubenswrapper[4676]: I0124 00:15:39.364259 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:15:39 crc kubenswrapper[4676]: I0124 00:15:39.364733 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:15:39 crc kubenswrapper[4676]: I0124 00:15:39.793612 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-lkqsz" Jan 24 00:15:40 crc kubenswrapper[4676]: I0124 00:15:40.073704 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:40 crc kubenswrapper[4676]: I0124 00:15:40.073783 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:40 crc kubenswrapper[4676]: I0124 00:15:40.082163 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:40 crc kubenswrapper[4676]: I0124 00:15:40.148836 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-646b8b7948-5mzrb" Jan 24 00:15:40 crc kubenswrapper[4676]: I0124 00:15:40.222977 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-g2smk"] Jan 24 00:15:49 crc kubenswrapper[4676]: I0124 00:15:49.718306 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-cpbrv" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.278326 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-g2smk" podUID="5cce043a-2f1b-4f48-967e-c48a00cfe1a6" containerName="console" containerID="cri-o://c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250" gracePeriod=15 Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.624666 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-g2smk_5cce043a-2f1b-4f48-967e-c48a00cfe1a6/console/0.log" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.624947 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.699077 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-serving-cert\") pod \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.699112 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6dpw\" (UniqueName: \"kubernetes.io/projected/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-kube-api-access-f6dpw\") pod \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.699163 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-service-ca\") pod \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.699243 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-oauth-config\") pod \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.699270 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-config\") pod \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.699288 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-oauth-serving-cert\") pod \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.699312 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-trusted-ca-bundle\") pod \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\" (UID: \"5cce043a-2f1b-4f48-967e-c48a00cfe1a6\") " Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.699879 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "5cce043a-2f1b-4f48-967e-c48a00cfe1a6" (UID: "5cce043a-2f1b-4f48-967e-c48a00cfe1a6"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.699889 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-service-ca" (OuterVolumeSpecName: "service-ca") pod "5cce043a-2f1b-4f48-967e-c48a00cfe1a6" (UID: "5cce043a-2f1b-4f48-967e-c48a00cfe1a6"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.700131 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-config" (OuterVolumeSpecName: "console-config") pod "5cce043a-2f1b-4f48-967e-c48a00cfe1a6" (UID: "5cce043a-2f1b-4f48-967e-c48a00cfe1a6"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.700287 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "5cce043a-2f1b-4f48-967e-c48a00cfe1a6" (UID: "5cce043a-2f1b-4f48-967e-c48a00cfe1a6"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.704066 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-kube-api-access-f6dpw" (OuterVolumeSpecName: "kube-api-access-f6dpw") pod "5cce043a-2f1b-4f48-967e-c48a00cfe1a6" (UID: "5cce043a-2f1b-4f48-967e-c48a00cfe1a6"). InnerVolumeSpecName "kube-api-access-f6dpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.704131 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "5cce043a-2f1b-4f48-967e-c48a00cfe1a6" (UID: "5cce043a-2f1b-4f48-967e-c48a00cfe1a6"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.707191 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "5cce043a-2f1b-4f48-967e-c48a00cfe1a6" (UID: "5cce043a-2f1b-4f48-967e-c48a00cfe1a6"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.801042 4676 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.801084 4676 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.801103 4676 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.801123 4676 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.801139 4676 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.801156 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6dpw\" (UniqueName: \"kubernetes.io/projected/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-kube-api-access-f6dpw\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.801173 4676 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5cce043a-2f1b-4f48-967e-c48a00cfe1a6-service-ca\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.893976 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j"] Jan 24 00:16:05 crc kubenswrapper[4676]: E0124 00:16:05.894288 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cce043a-2f1b-4f48-967e-c48a00cfe1a6" containerName="console" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.894307 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cce043a-2f1b-4f48-967e-c48a00cfe1a6" containerName="console" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.894849 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cce043a-2f1b-4f48-967e-c48a00cfe1a6" containerName="console" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.896070 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.898100 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.902093 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.902167 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7wz4\" (UniqueName: \"kubernetes.io/projected/3c111688-154b-47fa-8f89-6e48007b1fec-kube-api-access-p7wz4\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.902194 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:05 crc kubenswrapper[4676]: I0124 00:16:05.914590 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j"] Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.003409 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7wz4\" (UniqueName: \"kubernetes.io/projected/3c111688-154b-47fa-8f89-6e48007b1fec-kube-api-access-p7wz4\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.003651 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.003765 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.004291 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.004514 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.025262 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7wz4\" (UniqueName: \"kubernetes.io/projected/3c111688-154b-47fa-8f89-6e48007b1fec-kube-api-access-p7wz4\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.221636 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.312106 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-g2smk_5cce043a-2f1b-4f48-967e-c48a00cfe1a6/console/0.log" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.312169 4676 generic.go:334] "Generic (PLEG): container finished" podID="5cce043a-2f1b-4f48-967e-c48a00cfe1a6" containerID="c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250" exitCode=2 Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.312207 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g2smk" event={"ID":"5cce043a-2f1b-4f48-967e-c48a00cfe1a6","Type":"ContainerDied","Data":"c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250"} Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.312245 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-g2smk" event={"ID":"5cce043a-2f1b-4f48-967e-c48a00cfe1a6","Type":"ContainerDied","Data":"8150da0c7bf43ea052c8b9cb2dfefad3eaa9d5a8d1321ad37559f3b798f3d407"} Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.312287 4676 scope.go:117] "RemoveContainer" containerID="c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.312470 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-g2smk" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.332567 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-g2smk"] Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.340974 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-g2smk"] Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.350198 4676 scope.go:117] "RemoveContainer" containerID="c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250" Jan 24 00:16:06 crc kubenswrapper[4676]: E0124 00:16:06.351840 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250\": container with ID starting with c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250 not found: ID does not exist" containerID="c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.351884 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250"} err="failed to get container status \"c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250\": rpc error: code = NotFound desc = could not find container \"c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250\": container with ID starting with c3846591c5d0c38a4ca3a3f7394a46d77743449e795f56e0255610d9941b7250 not found: ID does not exist" Jan 24 00:16:06 crc kubenswrapper[4676]: E0124 00:16:06.414496 4676 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cce043a_2f1b_4f48_967e_c48a00cfe1a6.slice/crio-8150da0c7bf43ea052c8b9cb2dfefad3eaa9d5a8d1321ad37559f3b798f3d407\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cce043a_2f1b_4f48_967e_c48a00cfe1a6.slice\": RecentStats: unable to find data in memory cache]" Jan 24 00:16:06 crc kubenswrapper[4676]: I0124 00:16:06.486850 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j"] Jan 24 00:16:07 crc kubenswrapper[4676]: I0124 00:16:07.320423 4676 generic.go:334] "Generic (PLEG): container finished" podID="3c111688-154b-47fa-8f89-6e48007b1fec" containerID="a83c10df42bff0ddbfa91eb239302d1389b7d6db480c439fd5f94c85af065f2d" exitCode=0 Jan 24 00:16:07 crc kubenswrapper[4676]: I0124 00:16:07.320556 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" event={"ID":"3c111688-154b-47fa-8f89-6e48007b1fec","Type":"ContainerDied","Data":"a83c10df42bff0ddbfa91eb239302d1389b7d6db480c439fd5f94c85af065f2d"} Jan 24 00:16:07 crc kubenswrapper[4676]: I0124 00:16:07.324197 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" event={"ID":"3c111688-154b-47fa-8f89-6e48007b1fec","Type":"ContainerStarted","Data":"76a3eda93838b6bb0e5c18b77075ac901cd959dcd48b40594a45fdb7a357e745"} Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.246056 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nfz8s"] Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.248020 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.273698 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cce043a-2f1b-4f48-967e-c48a00cfe1a6" path="/var/lib/kubelet/pods/5cce043a-2f1b-4f48-967e-c48a00cfe1a6/volumes" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.274891 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nfz8s"] Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.444933 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-utilities\") pod \"redhat-operators-nfz8s\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.445017 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hnz4\" (UniqueName: \"kubernetes.io/projected/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-kube-api-access-9hnz4\") pod \"redhat-operators-nfz8s\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.445165 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-catalog-content\") pod \"redhat-operators-nfz8s\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.546336 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-catalog-content\") pod \"redhat-operators-nfz8s\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.546421 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-utilities\") pod \"redhat-operators-nfz8s\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.546462 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hnz4\" (UniqueName: \"kubernetes.io/projected/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-kube-api-access-9hnz4\") pod \"redhat-operators-nfz8s\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.547263 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-catalog-content\") pod \"redhat-operators-nfz8s\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.547586 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-utilities\") pod \"redhat-operators-nfz8s\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.566590 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hnz4\" (UniqueName: \"kubernetes.io/projected/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-kube-api-access-9hnz4\") pod \"redhat-operators-nfz8s\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.576932 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:08 crc kubenswrapper[4676]: I0124 00:16:08.787263 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nfz8s"] Jan 24 00:16:08 crc kubenswrapper[4676]: W0124 00:16:08.798803 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca7145d4_57a1_4ad4_acc8_b7ca02cfd407.slice/crio-8153314f2ab7211dbb0879d3369667b48e7906201af8f125d1bc6d0c80e1209a WatchSource:0}: Error finding container 8153314f2ab7211dbb0879d3369667b48e7906201af8f125d1bc6d0c80e1209a: Status 404 returned error can't find the container with id 8153314f2ab7211dbb0879d3369667b48e7906201af8f125d1bc6d0c80e1209a Jan 24 00:16:09 crc kubenswrapper[4676]: I0124 00:16:09.337984 4676 generic.go:334] "Generic (PLEG): container finished" podID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerID="2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9" exitCode=0 Jan 24 00:16:09 crc kubenswrapper[4676]: I0124 00:16:09.338131 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nfz8s" event={"ID":"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407","Type":"ContainerDied","Data":"2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9"} Jan 24 00:16:09 crc kubenswrapper[4676]: I0124 00:16:09.338500 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nfz8s" event={"ID":"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407","Type":"ContainerStarted","Data":"8153314f2ab7211dbb0879d3369667b48e7906201af8f125d1bc6d0c80e1209a"} Jan 24 00:16:09 crc kubenswrapper[4676]: I0124 00:16:09.364453 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:16:09 crc kubenswrapper[4676]: I0124 00:16:09.364515 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:16:11 crc kubenswrapper[4676]: I0124 00:16:11.310460 4676 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 24 00:16:12 crc kubenswrapper[4676]: I0124 00:16:12.357957 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nfz8s" event={"ID":"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407","Type":"ContainerStarted","Data":"ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c"} Jan 24 00:16:12 crc kubenswrapper[4676]: I0124 00:16:12.360220 4676 generic.go:334] "Generic (PLEG): container finished" podID="3c111688-154b-47fa-8f89-6e48007b1fec" containerID="38dee7997e1559ef2efa2b8cdca200f0488dba16c0166cda27d0639e2bae7090" exitCode=0 Jan 24 00:16:12 crc kubenswrapper[4676]: I0124 00:16:12.360277 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" event={"ID":"3c111688-154b-47fa-8f89-6e48007b1fec","Type":"ContainerDied","Data":"38dee7997e1559ef2efa2b8cdca200f0488dba16c0166cda27d0639e2bae7090"} Jan 24 00:16:13 crc kubenswrapper[4676]: I0124 00:16:13.370444 4676 generic.go:334] "Generic (PLEG): container finished" podID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerID="ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c" exitCode=0 Jan 24 00:16:13 crc kubenswrapper[4676]: I0124 00:16:13.370495 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nfz8s" event={"ID":"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407","Type":"ContainerDied","Data":"ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c"} Jan 24 00:16:13 crc kubenswrapper[4676]: I0124 00:16:13.375783 4676 generic.go:334] "Generic (PLEG): container finished" podID="3c111688-154b-47fa-8f89-6e48007b1fec" containerID="6e8cb074fed9adee705d124b9629ba61703c8cee8afacdcc25b520591cb5c008" exitCode=0 Jan 24 00:16:13 crc kubenswrapper[4676]: I0124 00:16:13.375845 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" event={"ID":"3c111688-154b-47fa-8f89-6e48007b1fec","Type":"ContainerDied","Data":"6e8cb074fed9adee705d124b9629ba61703c8cee8afacdcc25b520591cb5c008"} Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.383642 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nfz8s" event={"ID":"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407","Type":"ContainerStarted","Data":"d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005"} Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.406675 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nfz8s" podStartSLOduration=2.85597101 podStartE2EDuration="6.40665859s" podCreationTimestamp="2026-01-24 00:16:08 +0000 UTC" firstStartedPulling="2026-01-24 00:16:10.349212378 +0000 UTC m=+754.379183379" lastFinishedPulling="2026-01-24 00:16:13.899899958 +0000 UTC m=+757.929870959" observedRunningTime="2026-01-24 00:16:14.403128602 +0000 UTC m=+758.433099613" watchObservedRunningTime="2026-01-24 00:16:14.40665859 +0000 UTC m=+758.436629591" Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.625266 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.632141 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7wz4\" (UniqueName: \"kubernetes.io/projected/3c111688-154b-47fa-8f89-6e48007b1fec-kube-api-access-p7wz4\") pod \"3c111688-154b-47fa-8f89-6e48007b1fec\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.632203 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-bundle\") pod \"3c111688-154b-47fa-8f89-6e48007b1fec\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.632230 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-util\") pod \"3c111688-154b-47fa-8f89-6e48007b1fec\" (UID: \"3c111688-154b-47fa-8f89-6e48007b1fec\") " Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.633601 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-bundle" (OuterVolumeSpecName: "bundle") pod "3c111688-154b-47fa-8f89-6e48007b1fec" (UID: "3c111688-154b-47fa-8f89-6e48007b1fec"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.638732 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c111688-154b-47fa-8f89-6e48007b1fec-kube-api-access-p7wz4" (OuterVolumeSpecName: "kube-api-access-p7wz4") pod "3c111688-154b-47fa-8f89-6e48007b1fec" (UID: "3c111688-154b-47fa-8f89-6e48007b1fec"). InnerVolumeSpecName "kube-api-access-p7wz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.643395 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-util" (OuterVolumeSpecName: "util") pod "3c111688-154b-47fa-8f89-6e48007b1fec" (UID: "3c111688-154b-47fa-8f89-6e48007b1fec"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.733083 4676 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.733114 4676 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3c111688-154b-47fa-8f89-6e48007b1fec-util\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:14 crc kubenswrapper[4676]: I0124 00:16:14.733125 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7wz4\" (UniqueName: \"kubernetes.io/projected/3c111688-154b-47fa-8f89-6e48007b1fec-kube-api-access-p7wz4\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:15 crc kubenswrapper[4676]: I0124 00:16:15.393828 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" event={"ID":"3c111688-154b-47fa-8f89-6e48007b1fec","Type":"ContainerDied","Data":"76a3eda93838b6bb0e5c18b77075ac901cd959dcd48b40594a45fdb7a357e745"} Jan 24 00:16:15 crc kubenswrapper[4676]: I0124 00:16:15.394130 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76a3eda93838b6bb0e5c18b77075ac901cd959dcd48b40594a45fdb7a357e745" Jan 24 00:16:15 crc kubenswrapper[4676]: I0124 00:16:15.393908 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j" Jan 24 00:16:18 crc kubenswrapper[4676]: I0124 00:16:18.577493 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:18 crc kubenswrapper[4676]: I0124 00:16:18.577563 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:19 crc kubenswrapper[4676]: I0124 00:16:19.630781 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nfz8s" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerName="registry-server" probeResult="failure" output=< Jan 24 00:16:19 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:16:19 crc kubenswrapper[4676]: > Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.008950 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb"] Jan 24 00:16:24 crc kubenswrapper[4676]: E0124 00:16:24.009357 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c111688-154b-47fa-8f89-6e48007b1fec" containerName="util" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.009367 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c111688-154b-47fa-8f89-6e48007b1fec" containerName="util" Jan 24 00:16:24 crc kubenswrapper[4676]: E0124 00:16:24.009390 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c111688-154b-47fa-8f89-6e48007b1fec" containerName="pull" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.009396 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c111688-154b-47fa-8f89-6e48007b1fec" containerName="pull" Jan 24 00:16:24 crc kubenswrapper[4676]: E0124 00:16:24.009408 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c111688-154b-47fa-8f89-6e48007b1fec" containerName="extract" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.009414 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c111688-154b-47fa-8f89-6e48007b1fec" containerName="extract" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.009505 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c111688-154b-47fa-8f89-6e48007b1fec" containerName="extract" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.010033 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.012683 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.013062 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fgjmf" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.013812 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.013977 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.014105 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.040786 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb"] Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.149881 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrnlp\" (UniqueName: \"kubernetes.io/projected/59200cad-bd1a-472a-a1a1-adccb5211b21-kube-api-access-mrnlp\") pod \"metallb-operator-controller-manager-6868db74d6-qfhwb\" (UID: \"59200cad-bd1a-472a-a1a1-adccb5211b21\") " pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.149954 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/59200cad-bd1a-472a-a1a1-adccb5211b21-webhook-cert\") pod \"metallb-operator-controller-manager-6868db74d6-qfhwb\" (UID: \"59200cad-bd1a-472a-a1a1-adccb5211b21\") " pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.150092 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/59200cad-bd1a-472a-a1a1-adccb5211b21-apiservice-cert\") pod \"metallb-operator-controller-manager-6868db74d6-qfhwb\" (UID: \"59200cad-bd1a-472a-a1a1-adccb5211b21\") " pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.246574 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5"] Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.247449 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.250519 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.250941 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrnlp\" (UniqueName: \"kubernetes.io/projected/59200cad-bd1a-472a-a1a1-adccb5211b21-kube-api-access-mrnlp\") pod \"metallb-operator-controller-manager-6868db74d6-qfhwb\" (UID: \"59200cad-bd1a-472a-a1a1-adccb5211b21\") " pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.250991 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/59200cad-bd1a-472a-a1a1-adccb5211b21-webhook-cert\") pod \"metallb-operator-controller-manager-6868db74d6-qfhwb\" (UID: \"59200cad-bd1a-472a-a1a1-adccb5211b21\") " pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.251018 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/59200cad-bd1a-472a-a1a1-adccb5211b21-apiservice-cert\") pod \"metallb-operator-controller-manager-6868db74d6-qfhwb\" (UID: \"59200cad-bd1a-472a-a1a1-adccb5211b21\") " pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.252945 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.253026 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-ckwmv" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.257179 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/59200cad-bd1a-472a-a1a1-adccb5211b21-apiservice-cert\") pod \"metallb-operator-controller-manager-6868db74d6-qfhwb\" (UID: \"59200cad-bd1a-472a-a1a1-adccb5211b21\") " pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.266247 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/59200cad-bd1a-472a-a1a1-adccb5211b21-webhook-cert\") pod \"metallb-operator-controller-manager-6868db74d6-qfhwb\" (UID: \"59200cad-bd1a-472a-a1a1-adccb5211b21\") " pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.288080 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrnlp\" (UniqueName: \"kubernetes.io/projected/59200cad-bd1a-472a-a1a1-adccb5211b21-kube-api-access-mrnlp\") pod \"metallb-operator-controller-manager-6868db74d6-qfhwb\" (UID: \"59200cad-bd1a-472a-a1a1-adccb5211b21\") " pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.310554 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5"] Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.323753 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.352217 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvj86\" (UniqueName: \"kubernetes.io/projected/adfe0dac-5ac5-44b8-97db-088d1ac83d34-kube-api-access-tvj86\") pod \"metallb-operator-webhook-server-6dbffff8f5-cb5l5\" (UID: \"adfe0dac-5ac5-44b8-97db-088d1ac83d34\") " pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.353025 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/adfe0dac-5ac5-44b8-97db-088d1ac83d34-apiservice-cert\") pod \"metallb-operator-webhook-server-6dbffff8f5-cb5l5\" (UID: \"adfe0dac-5ac5-44b8-97db-088d1ac83d34\") " pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.353211 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/adfe0dac-5ac5-44b8-97db-088d1ac83d34-webhook-cert\") pod \"metallb-operator-webhook-server-6dbffff8f5-cb5l5\" (UID: \"adfe0dac-5ac5-44b8-97db-088d1ac83d34\") " pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.458825 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvj86\" (UniqueName: \"kubernetes.io/projected/adfe0dac-5ac5-44b8-97db-088d1ac83d34-kube-api-access-tvj86\") pod \"metallb-operator-webhook-server-6dbffff8f5-cb5l5\" (UID: \"adfe0dac-5ac5-44b8-97db-088d1ac83d34\") " pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.458863 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/adfe0dac-5ac5-44b8-97db-088d1ac83d34-apiservice-cert\") pod \"metallb-operator-webhook-server-6dbffff8f5-cb5l5\" (UID: \"adfe0dac-5ac5-44b8-97db-088d1ac83d34\") " pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.458931 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/adfe0dac-5ac5-44b8-97db-088d1ac83d34-webhook-cert\") pod \"metallb-operator-webhook-server-6dbffff8f5-cb5l5\" (UID: \"adfe0dac-5ac5-44b8-97db-088d1ac83d34\") " pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.462946 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/adfe0dac-5ac5-44b8-97db-088d1ac83d34-apiservice-cert\") pod \"metallb-operator-webhook-server-6dbffff8f5-cb5l5\" (UID: \"adfe0dac-5ac5-44b8-97db-088d1ac83d34\") " pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.463337 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/adfe0dac-5ac5-44b8-97db-088d1ac83d34-webhook-cert\") pod \"metallb-operator-webhook-server-6dbffff8f5-cb5l5\" (UID: \"adfe0dac-5ac5-44b8-97db-088d1ac83d34\") " pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.477665 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvj86\" (UniqueName: \"kubernetes.io/projected/adfe0dac-5ac5-44b8-97db-088d1ac83d34-kube-api-access-tvj86\") pod \"metallb-operator-webhook-server-6dbffff8f5-cb5l5\" (UID: \"adfe0dac-5ac5-44b8-97db-088d1ac83d34\") " pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.592680 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.642652 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb"] Jan 24 00:16:24 crc kubenswrapper[4676]: I0124 00:16:24.924935 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5"] Jan 24 00:16:24 crc kubenswrapper[4676]: W0124 00:16:24.933169 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadfe0dac_5ac5_44b8_97db_088d1ac83d34.slice/crio-50d12f22e386ef9c2b41817b928dd900066a8e50105761cc692396dae78c8ec4 WatchSource:0}: Error finding container 50d12f22e386ef9c2b41817b928dd900066a8e50105761cc692396dae78c8ec4: Status 404 returned error can't find the container with id 50d12f22e386ef9c2b41817b928dd900066a8e50105761cc692396dae78c8ec4 Jan 24 00:16:25 crc kubenswrapper[4676]: I0124 00:16:25.446475 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" event={"ID":"59200cad-bd1a-472a-a1a1-adccb5211b21","Type":"ContainerStarted","Data":"0cb2a0d9fb39b905916b5fcc5105590a8bc9706a24b1d330dea1ede36464be07"} Jan 24 00:16:25 crc kubenswrapper[4676]: I0124 00:16:25.448792 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" event={"ID":"adfe0dac-5ac5-44b8-97db-088d1ac83d34","Type":"ContainerStarted","Data":"50d12f22e386ef9c2b41817b928dd900066a8e50105761cc692396dae78c8ec4"} Jan 24 00:16:28 crc kubenswrapper[4676]: I0124 00:16:28.624574 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:28 crc kubenswrapper[4676]: I0124 00:16:28.676158 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:28 crc kubenswrapper[4676]: I0124 00:16:28.856525 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nfz8s"] Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.475534 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" event={"ID":"59200cad-bd1a-472a-a1a1-adccb5211b21","Type":"ContainerStarted","Data":"aa93327a37132aaaf13467f36f40f5999ca32c6929f4734fba4815d29bef0ab0"} Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.476329 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.477276 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" event={"ID":"adfe0dac-5ac5-44b8-97db-088d1ac83d34","Type":"ContainerStarted","Data":"a4a1604634fabd24c3814d8510a89d532d058c19481ab88cd0b114d36f236462"} Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.477530 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nfz8s" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerName="registry-server" containerID="cri-o://d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005" gracePeriod=2 Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.516689 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" podStartSLOduration=2.030895947 podStartE2EDuration="7.516670154s" podCreationTimestamp="2026-01-24 00:16:23 +0000 UTC" firstStartedPulling="2026-01-24 00:16:24.686541083 +0000 UTC m=+768.716512084" lastFinishedPulling="2026-01-24 00:16:30.17231529 +0000 UTC m=+774.202286291" observedRunningTime="2026-01-24 00:16:30.513046613 +0000 UTC m=+774.543017614" watchObservedRunningTime="2026-01-24 00:16:30.516670154 +0000 UTC m=+774.546641155" Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.539338 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" podStartSLOduration=1.284091543 podStartE2EDuration="6.539322397s" podCreationTimestamp="2026-01-24 00:16:24 +0000 UTC" firstStartedPulling="2026-01-24 00:16:24.936477379 +0000 UTC m=+768.966448380" lastFinishedPulling="2026-01-24 00:16:30.191708233 +0000 UTC m=+774.221679234" observedRunningTime="2026-01-24 00:16:30.537861902 +0000 UTC m=+774.567832903" watchObservedRunningTime="2026-01-24 00:16:30.539322397 +0000 UTC m=+774.569293398" Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.844912 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.956347 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-catalog-content\") pod \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.956529 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-utilities\") pod \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.956589 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hnz4\" (UniqueName: \"kubernetes.io/projected/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-kube-api-access-9hnz4\") pod \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\" (UID: \"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407\") " Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.958066 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-utilities" (OuterVolumeSpecName: "utilities") pod "ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" (UID: "ca7145d4-57a1-4ad4-acc8-b7ca02cfd407"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:16:30 crc kubenswrapper[4676]: I0124 00:16:30.967289 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-kube-api-access-9hnz4" (OuterVolumeSpecName: "kube-api-access-9hnz4") pod "ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" (UID: "ca7145d4-57a1-4ad4-acc8-b7ca02cfd407"). InnerVolumeSpecName "kube-api-access-9hnz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.057789 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.057829 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hnz4\" (UniqueName: \"kubernetes.io/projected/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-kube-api-access-9hnz4\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.100233 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" (UID: "ca7145d4-57a1-4ad4-acc8-b7ca02cfd407"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.158568 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.493987 4676 generic.go:334] "Generic (PLEG): container finished" podID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerID="d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005" exitCode=0 Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.494091 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nfz8s" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.494187 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nfz8s" event={"ID":"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407","Type":"ContainerDied","Data":"d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005"} Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.495186 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nfz8s" event={"ID":"ca7145d4-57a1-4ad4-acc8-b7ca02cfd407","Type":"ContainerDied","Data":"8153314f2ab7211dbb0879d3369667b48e7906201af8f125d1bc6d0c80e1209a"} Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.495225 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.495258 4676 scope.go:117] "RemoveContainer" containerID="d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.525594 4676 scope.go:117] "RemoveContainer" containerID="ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.525656 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nfz8s"] Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.535155 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nfz8s"] Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.549372 4676 scope.go:117] "RemoveContainer" containerID="2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.573684 4676 scope.go:117] "RemoveContainer" containerID="d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005" Jan 24 00:16:31 crc kubenswrapper[4676]: E0124 00:16:31.574181 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005\": container with ID starting with d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005 not found: ID does not exist" containerID="d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.574219 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005"} err="failed to get container status \"d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005\": rpc error: code = NotFound desc = could not find container \"d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005\": container with ID starting with d1d85267432a4570db782b58a91df6a8201f6bad193dc5837dfdbc50d8377005 not found: ID does not exist" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.574245 4676 scope.go:117] "RemoveContainer" containerID="ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c" Jan 24 00:16:31 crc kubenswrapper[4676]: E0124 00:16:31.574918 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c\": container with ID starting with ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c not found: ID does not exist" containerID="ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.574944 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c"} err="failed to get container status \"ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c\": rpc error: code = NotFound desc = could not find container \"ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c\": container with ID starting with ab3b8aade9c01793933d385fb272beb855e6bd497784c470d8ff64768ede8f1c not found: ID does not exist" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.574962 4676 scope.go:117] "RemoveContainer" containerID="2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9" Jan 24 00:16:31 crc kubenswrapper[4676]: E0124 00:16:31.575307 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9\": container with ID starting with 2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9 not found: ID does not exist" containerID="2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9" Jan 24 00:16:31 crc kubenswrapper[4676]: I0124 00:16:31.575355 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9"} err="failed to get container status \"2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9\": rpc error: code = NotFound desc = could not find container \"2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9\": container with ID starting with 2687c4aebd5cf8c6cdb52a49834465bc2ff9fdc1624a9f7ea9eb31f3d327a7a9 not found: ID does not exist" Jan 24 00:16:32 crc kubenswrapper[4676]: I0124 00:16:32.264319 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" path="/var/lib/kubelet/pods/ca7145d4-57a1-4ad4-acc8-b7ca02cfd407/volumes" Jan 24 00:16:39 crc kubenswrapper[4676]: I0124 00:16:39.364334 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:16:39 crc kubenswrapper[4676]: I0124 00:16:39.364804 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:16:39 crc kubenswrapper[4676]: I0124 00:16:39.364849 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:16:39 crc kubenswrapper[4676]: I0124 00:16:39.365392 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"daf4dfd81dc7faee8c5a37cce872ffde5731f2d91708788dd42d2993fec18ba6"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:16:39 crc kubenswrapper[4676]: I0124 00:16:39.365440 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://daf4dfd81dc7faee8c5a37cce872ffde5731f2d91708788dd42d2993fec18ba6" gracePeriod=600 Jan 24 00:16:40 crc kubenswrapper[4676]: I0124 00:16:40.544345 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="daf4dfd81dc7faee8c5a37cce872ffde5731f2d91708788dd42d2993fec18ba6" exitCode=0 Jan 24 00:16:40 crc kubenswrapper[4676]: I0124 00:16:40.544409 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"daf4dfd81dc7faee8c5a37cce872ffde5731f2d91708788dd42d2993fec18ba6"} Jan 24 00:16:40 crc kubenswrapper[4676]: I0124 00:16:40.544683 4676 scope.go:117] "RemoveContainer" containerID="51936651fd255c79ecefd9c199d1a0336083a72c956e86d582874060fc907470" Jan 24 00:16:41 crc kubenswrapper[4676]: I0124 00:16:41.566858 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"2687110039e3aba350a72ed3647bbafb008d22f301a8b50baa7159c6eca5ba33"} Jan 24 00:16:44 crc kubenswrapper[4676]: I0124 00:16:44.601148 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6dbffff8f5-cb5l5" Jan 24 00:17:04 crc kubenswrapper[4676]: I0124 00:17:04.328369 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6868db74d6-qfhwb" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.276133 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-szpqt"] Jan 24 00:17:05 crc kubenswrapper[4676]: E0124 00:17:05.276430 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerName="registry-server" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.276448 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerName="registry-server" Jan 24 00:17:05 crc kubenswrapper[4676]: E0124 00:17:05.276458 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerName="extract-content" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.276465 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerName="extract-content" Jan 24 00:17:05 crc kubenswrapper[4676]: E0124 00:17:05.276482 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerName="extract-utilities" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.276488 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerName="extract-utilities" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.276592 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca7145d4-57a1-4ad4-acc8-b7ca02cfd407" containerName="registry-server" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.278405 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.281572 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59"] Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.282499 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.282967 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.287114 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.287610 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-98z5t" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.296631 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59"] Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.298030 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.389284 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-zh6td"] Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.390139 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.393013 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.393191 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-cbd85" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.393552 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.393874 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.402477 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-reloader\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.402530 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-frr-conf\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.402586 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b5c2\" (UniqueName: \"kubernetes.io/projected/8a7798b0-f97f-4804-a406-22ecc8b45677-kube-api-access-7b5c2\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.402614 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftkn7\" (UniqueName: \"kubernetes.io/projected/2ce22d83-ee4f-4ad7-8882-b876d4ed52a2-kube-api-access-ftkn7\") pod \"frr-k8s-webhook-server-7df86c4f6c-s7b59\" (UID: \"2ce22d83-ee4f-4ad7-8882-b876d4ed52a2\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.402637 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8a7798b0-f97f-4804-a406-22ecc8b45677-frr-startup\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.402661 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ce22d83-ee4f-4ad7-8882-b876d4ed52a2-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-s7b59\" (UID: \"2ce22d83-ee4f-4ad7-8882-b876d4ed52a2\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.402675 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-frr-sockets\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.402690 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-metrics\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.402727 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a7798b0-f97f-4804-a406-22ecc8b45677-metrics-certs\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.408399 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-jqk4f"] Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.409139 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.410495 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.430685 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-jqk4f"] Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.503817 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8a7798b0-f97f-4804-a406-22ecc8b45677-frr-startup\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.503854 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ce22d83-ee4f-4ad7-8882-b876d4ed52a2-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-s7b59\" (UID: \"2ce22d83-ee4f-4ad7-8882-b876d4ed52a2\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.503872 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-frr-sockets\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.503891 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-metrics\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.503923 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a7798b0-f97f-4804-a406-22ecc8b45677-metrics-certs\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.503947 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5517856-fac7-4312-ab46-86bbd5c1282d-cert\") pod \"controller-6968d8fdc4-jqk4f\" (UID: \"f5517856-fac7-4312-ab46-86bbd5c1282d\") " pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.503998 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwl55\" (UniqueName: \"kubernetes.io/projected/7da3019e-01de-4671-a78f-6c0d2e57fde3-kube-api-access-qwl55\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504018 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-reloader\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: E0124 00:17:05.504046 4676 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504068 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-metrics-certs\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: E0124 00:17:05.504117 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8a7798b0-f97f-4804-a406-22ecc8b45677-metrics-certs podName:8a7798b0-f97f-4804-a406-22ecc8b45677 nodeName:}" failed. No retries permitted until 2026-01-24 00:17:06.004099798 +0000 UTC m=+810.034070799 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8a7798b0-f97f-4804-a406-22ecc8b45677-metrics-certs") pod "frr-k8s-szpqt" (UID: "8a7798b0-f97f-4804-a406-22ecc8b45677") : secret "frr-k8s-certs-secret" not found Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504148 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5517856-fac7-4312-ab46-86bbd5c1282d-metrics-certs\") pod \"controller-6968d8fdc4-jqk4f\" (UID: \"f5517856-fac7-4312-ab46-86bbd5c1282d\") " pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504172 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-frr-conf\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504194 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7da3019e-01de-4671-a78f-6c0d2e57fde3-metallb-excludel2\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504220 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b5c2\" (UniqueName: \"kubernetes.io/projected/8a7798b0-f97f-4804-a406-22ecc8b45677-kube-api-access-7b5c2\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504243 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrkgb\" (UniqueName: \"kubernetes.io/projected/f5517856-fac7-4312-ab46-86bbd5c1282d-kube-api-access-rrkgb\") pod \"controller-6968d8fdc4-jqk4f\" (UID: \"f5517856-fac7-4312-ab46-86bbd5c1282d\") " pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504267 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftkn7\" (UniqueName: \"kubernetes.io/projected/2ce22d83-ee4f-4ad7-8882-b876d4ed52a2-kube-api-access-ftkn7\") pod \"frr-k8s-webhook-server-7df86c4f6c-s7b59\" (UID: \"2ce22d83-ee4f-4ad7-8882-b876d4ed52a2\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504293 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-memberlist\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504369 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-metrics\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504542 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-frr-conf\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504748 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-frr-sockets\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504771 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8a7798b0-f97f-4804-a406-22ecc8b45677-frr-startup\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.504888 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8a7798b0-f97f-4804-a406-22ecc8b45677-reloader\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.522312 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ce22d83-ee4f-4ad7-8882-b876d4ed52a2-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-s7b59\" (UID: \"2ce22d83-ee4f-4ad7-8882-b876d4ed52a2\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.524841 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b5c2\" (UniqueName: \"kubernetes.io/projected/8a7798b0-f97f-4804-a406-22ecc8b45677-kube-api-access-7b5c2\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.525416 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftkn7\" (UniqueName: \"kubernetes.io/projected/2ce22d83-ee4f-4ad7-8882-b876d4ed52a2-kube-api-access-ftkn7\") pod \"frr-k8s-webhook-server-7df86c4f6c-s7b59\" (UID: \"2ce22d83-ee4f-4ad7-8882-b876d4ed52a2\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.602494 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.604834 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5517856-fac7-4312-ab46-86bbd5c1282d-metrics-certs\") pod \"controller-6968d8fdc4-jqk4f\" (UID: \"f5517856-fac7-4312-ab46-86bbd5c1282d\") " pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.604869 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7da3019e-01de-4671-a78f-6c0d2e57fde3-metallb-excludel2\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.604900 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrkgb\" (UniqueName: \"kubernetes.io/projected/f5517856-fac7-4312-ab46-86bbd5c1282d-kube-api-access-rrkgb\") pod \"controller-6968d8fdc4-jqk4f\" (UID: \"f5517856-fac7-4312-ab46-86bbd5c1282d\") " pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.604922 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-memberlist\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.604974 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5517856-fac7-4312-ab46-86bbd5c1282d-cert\") pod \"controller-6968d8fdc4-jqk4f\" (UID: \"f5517856-fac7-4312-ab46-86bbd5c1282d\") " pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.604992 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwl55\" (UniqueName: \"kubernetes.io/projected/7da3019e-01de-4671-a78f-6c0d2e57fde3-kube-api-access-qwl55\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.605012 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-metrics-certs\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.606699 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7da3019e-01de-4671-a78f-6c0d2e57fde3-metallb-excludel2\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: E0124 00:17:05.605610 4676 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 24 00:17:05 crc kubenswrapper[4676]: E0124 00:17:05.606813 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-memberlist podName:7da3019e-01de-4671-a78f-6c0d2e57fde3 nodeName:}" failed. No retries permitted until 2026-01-24 00:17:06.10678807 +0000 UTC m=+810.136759071 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-memberlist") pod "speaker-zh6td" (UID: "7da3019e-01de-4671-a78f-6c0d2e57fde3") : secret "metallb-memberlist" not found Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.607700 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-metrics-certs\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.613479 4676 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.633445 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwl55\" (UniqueName: \"kubernetes.io/projected/7da3019e-01de-4671-a78f-6c0d2e57fde3-kube-api-access-qwl55\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.635573 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrkgb\" (UniqueName: \"kubernetes.io/projected/f5517856-fac7-4312-ab46-86bbd5c1282d-kube-api-access-rrkgb\") pod \"controller-6968d8fdc4-jqk4f\" (UID: \"f5517856-fac7-4312-ab46-86bbd5c1282d\") " pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.637869 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f5517856-fac7-4312-ab46-86bbd5c1282d-metrics-certs\") pod \"controller-6968d8fdc4-jqk4f\" (UID: \"f5517856-fac7-4312-ab46-86bbd5c1282d\") " pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.638417 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f5517856-fac7-4312-ab46-86bbd5c1282d-cert\") pod \"controller-6968d8fdc4-jqk4f\" (UID: \"f5517856-fac7-4312-ab46-86bbd5c1282d\") " pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.720748 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.839368 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59"] Jan 24 00:17:05 crc kubenswrapper[4676]: I0124 00:17:05.897880 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-jqk4f"] Jan 24 00:17:05 crc kubenswrapper[4676]: W0124 00:17:05.902574 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5517856_fac7_4312_ab46_86bbd5c1282d.slice/crio-7067f52856da71bef338b31656e506d7ff1da9055f6a009db9a859bfeee222b0 WatchSource:0}: Error finding container 7067f52856da71bef338b31656e506d7ff1da9055f6a009db9a859bfeee222b0: Status 404 returned error can't find the container with id 7067f52856da71bef338b31656e506d7ff1da9055f6a009db9a859bfeee222b0 Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.021187 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a7798b0-f97f-4804-a406-22ecc8b45677-metrics-certs\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.026306 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8a7798b0-f97f-4804-a406-22ecc8b45677-metrics-certs\") pod \"frr-k8s-szpqt\" (UID: \"8a7798b0-f97f-4804-a406-22ecc8b45677\") " pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.121911 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-memberlist\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:06 crc kubenswrapper[4676]: E0124 00:17:06.122069 4676 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 24 00:17:06 crc kubenswrapper[4676]: E0124 00:17:06.122138 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-memberlist podName:7da3019e-01de-4671-a78f-6c0d2e57fde3 nodeName:}" failed. No retries permitted until 2026-01-24 00:17:07.122118644 +0000 UTC m=+811.152089665 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-memberlist") pod "speaker-zh6td" (UID: "7da3019e-01de-4671-a78f-6c0d2e57fde3") : secret "metallb-memberlist" not found Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.197817 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.731877 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" event={"ID":"2ce22d83-ee4f-4ad7-8882-b876d4ed52a2","Type":"ContainerStarted","Data":"a334528e9fcce16672cb9c179bec89136fc19677c417e99af539b99acfae4ef1"} Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.733498 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-jqk4f" event={"ID":"f5517856-fac7-4312-ab46-86bbd5c1282d","Type":"ContainerStarted","Data":"2c5cc057854c659ba305a531384a04657496d8ab3c63ea3c510ee6ee420d51fe"} Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.733523 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-jqk4f" event={"ID":"f5517856-fac7-4312-ab46-86bbd5c1282d","Type":"ContainerStarted","Data":"68aec8e91b27920e6ce721982604cb53f4298058a221854b41ee0607d62c9cc5"} Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.733537 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-jqk4f" event={"ID":"f5517856-fac7-4312-ab46-86bbd5c1282d","Type":"ContainerStarted","Data":"7067f52856da71bef338b31656e506d7ff1da9055f6a009db9a859bfeee222b0"} Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.733583 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.734619 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerStarted","Data":"a476192e18a4d4b2e0b1eb57b0acf4c20d24261e27c719e71607eef42208dee8"} Jan 24 00:17:06 crc kubenswrapper[4676]: I0124 00:17:06.751926 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-jqk4f" podStartSLOduration=1.75190868 podStartE2EDuration="1.75190868s" podCreationTimestamp="2026-01-24 00:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:17:06.74896235 +0000 UTC m=+810.778933351" watchObservedRunningTime="2026-01-24 00:17:06.75190868 +0000 UTC m=+810.781879681" Jan 24 00:17:07 crc kubenswrapper[4676]: I0124 00:17:07.139080 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-memberlist\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:07 crc kubenswrapper[4676]: I0124 00:17:07.150629 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7da3019e-01de-4671-a78f-6c0d2e57fde3-memberlist\") pod \"speaker-zh6td\" (UID: \"7da3019e-01de-4671-a78f-6c0d2e57fde3\") " pod="metallb-system/speaker-zh6td" Jan 24 00:17:07 crc kubenswrapper[4676]: I0124 00:17:07.202545 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zh6td" Jan 24 00:17:07 crc kubenswrapper[4676]: I0124 00:17:07.741800 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zh6td" event={"ID":"7da3019e-01de-4671-a78f-6c0d2e57fde3","Type":"ContainerStarted","Data":"b5974bf994f467296463e48aa1ec39d04a14897f0084f65a861afbe278a93e9b"} Jan 24 00:17:07 crc kubenswrapper[4676]: I0124 00:17:07.742116 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zh6td" event={"ID":"7da3019e-01de-4671-a78f-6c0d2e57fde3","Type":"ContainerStarted","Data":"f25a970d7f4afe68b1d16f8a6f98e4e45811ab615364d8b0c2c28c0b4488306f"} Jan 24 00:17:08 crc kubenswrapper[4676]: I0124 00:17:08.793930 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zh6td" event={"ID":"7da3019e-01de-4671-a78f-6c0d2e57fde3","Type":"ContainerStarted","Data":"54b95f7fa5ad58bcb9fb952d4e0d97ba3049a39c032a908cae1f94b4d39a039a"} Jan 24 00:17:08 crc kubenswrapper[4676]: I0124 00:17:08.795014 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-zh6td" Jan 24 00:17:08 crc kubenswrapper[4676]: I0124 00:17:08.844740 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-zh6td" podStartSLOduration=3.8447223619999997 podStartE2EDuration="3.844722362s" podCreationTimestamp="2026-01-24 00:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:17:08.840692578 +0000 UTC m=+812.870663579" watchObservedRunningTime="2026-01-24 00:17:08.844722362 +0000 UTC m=+812.874693363" Jan 24 00:17:14 crc kubenswrapper[4676]: I0124 00:17:14.825659 4676 generic.go:334] "Generic (PLEG): container finished" podID="8a7798b0-f97f-4804-a406-22ecc8b45677" containerID="0f872ee63e1cdba2e1e4761f82b3a98b3f10875ae88354fd9c9a65bcdf49ada5" exitCode=0 Jan 24 00:17:14 crc kubenswrapper[4676]: I0124 00:17:14.825757 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerDied","Data":"0f872ee63e1cdba2e1e4761f82b3a98b3f10875ae88354fd9c9a65bcdf49ada5"} Jan 24 00:17:14 crc kubenswrapper[4676]: I0124 00:17:14.828887 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" event={"ID":"2ce22d83-ee4f-4ad7-8882-b876d4ed52a2","Type":"ContainerStarted","Data":"9468eea7ba989d37d33d89820231269843b9aa5eb2e025ae0bc27dc696636687"} Jan 24 00:17:14 crc kubenswrapper[4676]: I0124 00:17:14.829006 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:14 crc kubenswrapper[4676]: I0124 00:17:14.876915 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" podStartSLOduration=1.765514596 podStartE2EDuration="9.876898442s" podCreationTimestamp="2026-01-24 00:17:05 +0000 UTC" firstStartedPulling="2026-01-24 00:17:05.850059932 +0000 UTC m=+809.880030933" lastFinishedPulling="2026-01-24 00:17:13.961443768 +0000 UTC m=+817.991414779" observedRunningTime="2026-01-24 00:17:14.875897322 +0000 UTC m=+818.905868323" watchObservedRunningTime="2026-01-24 00:17:14.876898442 +0000 UTC m=+818.906869443" Jan 24 00:17:15 crc kubenswrapper[4676]: I0124 00:17:15.845770 4676 generic.go:334] "Generic (PLEG): container finished" podID="8a7798b0-f97f-4804-a406-22ecc8b45677" containerID="985d356157001ea96677721ef7091464e7dcd5614907239f9ff5a6b869083e2c" exitCode=0 Jan 24 00:17:15 crc kubenswrapper[4676]: I0124 00:17:15.845992 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerDied","Data":"985d356157001ea96677721ef7091464e7dcd5614907239f9ff5a6b869083e2c"} Jan 24 00:17:16 crc kubenswrapper[4676]: I0124 00:17:16.854046 4676 generic.go:334] "Generic (PLEG): container finished" podID="8a7798b0-f97f-4804-a406-22ecc8b45677" containerID="45b168d21f1a0451303b67d18b62f8d1c2eb635d47fbb9c410ec73ac7826281e" exitCode=0 Jan 24 00:17:16 crc kubenswrapper[4676]: I0124 00:17:16.854096 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerDied","Data":"45b168d21f1a0451303b67d18b62f8d1c2eb635d47fbb9c410ec73ac7826281e"} Jan 24 00:17:17 crc kubenswrapper[4676]: I0124 00:17:17.207833 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-zh6td" Jan 24 00:17:17 crc kubenswrapper[4676]: I0124 00:17:17.865900 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerStarted","Data":"639934f83046a208ac6c36d423b7b7a8a2fc9ac7a06f9d72cd3f40649b7625a5"} Jan 24 00:17:18 crc kubenswrapper[4676]: I0124 00:17:18.877180 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerStarted","Data":"89222040cea7839f9811fefb567adfa49b75cbc63cb29f34599ec76da946001d"} Jan 24 00:17:18 crc kubenswrapper[4676]: I0124 00:17:18.877512 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:18 crc kubenswrapper[4676]: I0124 00:17:18.877527 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerStarted","Data":"32cddb0f809234929d53e50a67f6ddf97e1ee1e6251d8a1c2347acefe6889e3d"} Jan 24 00:17:18 crc kubenswrapper[4676]: I0124 00:17:18.877540 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerStarted","Data":"657915650d3d7e15f1f43806bf755675fc8ae3f14e29e2da6a3a5c061c6a364b"} Jan 24 00:17:18 crc kubenswrapper[4676]: I0124 00:17:18.877552 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerStarted","Data":"ebfd023ac87216fefc6a03fae974686e76d94aa90ae27e027fffdfe8da7ac3a9"} Jan 24 00:17:18 crc kubenswrapper[4676]: I0124 00:17:18.877565 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-szpqt" event={"ID":"8a7798b0-f97f-4804-a406-22ecc8b45677","Type":"ContainerStarted","Data":"a2e58590b3d95ad6600c527f73595ba81f0848ac3129fd5349dd5406a3270855"} Jan 24 00:17:18 crc kubenswrapper[4676]: I0124 00:17:18.903964 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-szpqt" podStartSLOduration=6.241574195 podStartE2EDuration="13.903945115s" podCreationTimestamp="2026-01-24 00:17:05 +0000 UTC" firstStartedPulling="2026-01-24 00:17:06.322812944 +0000 UTC m=+810.352783945" lastFinishedPulling="2026-01-24 00:17:13.985183854 +0000 UTC m=+818.015154865" observedRunningTime="2026-01-24 00:17:18.898804137 +0000 UTC m=+822.928775138" watchObservedRunningTime="2026-01-24 00:17:18.903945115 +0000 UTC m=+822.933916136" Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.212251 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-6ntrt"] Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.213610 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6ntrt" Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.216629 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.217322 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-jsvzn" Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.221173 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.233648 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6ntrt"] Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.337199 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4vfp\" (UniqueName: \"kubernetes.io/projected/5425e3ea-c929-4943-bd14-d4e193992052-kube-api-access-r4vfp\") pod \"openstack-operator-index-6ntrt\" (UID: \"5425e3ea-c929-4943-bd14-d4e193992052\") " pod="openstack-operators/openstack-operator-index-6ntrt" Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.439304 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4vfp\" (UniqueName: \"kubernetes.io/projected/5425e3ea-c929-4943-bd14-d4e193992052-kube-api-access-r4vfp\") pod \"openstack-operator-index-6ntrt\" (UID: \"5425e3ea-c929-4943-bd14-d4e193992052\") " pod="openstack-operators/openstack-operator-index-6ntrt" Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.459212 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4vfp\" (UniqueName: \"kubernetes.io/projected/5425e3ea-c929-4943-bd14-d4e193992052-kube-api-access-r4vfp\") pod \"openstack-operator-index-6ntrt\" (UID: \"5425e3ea-c929-4943-bd14-d4e193992052\") " pod="openstack-operators/openstack-operator-index-6ntrt" Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.533641 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6ntrt" Jan 24 00:17:20 crc kubenswrapper[4676]: I0124 00:17:20.964987 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6ntrt"] Jan 24 00:17:21 crc kubenswrapper[4676]: I0124 00:17:21.198442 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:21 crc kubenswrapper[4676]: I0124 00:17:21.245917 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:21 crc kubenswrapper[4676]: I0124 00:17:21.893708 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6ntrt" event={"ID":"5425e3ea-c929-4943-bd14-d4e193992052","Type":"ContainerStarted","Data":"ee09d2706d4eadeb85bdfce3c04bc857b26be38730265e45634501b3005889aa"} Jan 24 00:17:23 crc kubenswrapper[4676]: I0124 00:17:23.908625 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6ntrt" event={"ID":"5425e3ea-c929-4943-bd14-d4e193992052","Type":"ContainerStarted","Data":"7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d"} Jan 24 00:17:23 crc kubenswrapper[4676]: I0124 00:17:23.928102 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-6ntrt" podStartSLOduration=1.600803535 podStartE2EDuration="3.928078419s" podCreationTimestamp="2026-01-24 00:17:20 +0000 UTC" firstStartedPulling="2026-01-24 00:17:20.9749779 +0000 UTC m=+825.004948911" lastFinishedPulling="2026-01-24 00:17:23.302252794 +0000 UTC m=+827.332223795" observedRunningTime="2026-01-24 00:17:23.925763398 +0000 UTC m=+827.955734399" watchObservedRunningTime="2026-01-24 00:17:23.928078419 +0000 UTC m=+827.958049460" Jan 24 00:17:24 crc kubenswrapper[4676]: I0124 00:17:24.388562 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6ntrt"] Jan 24 00:17:24 crc kubenswrapper[4676]: I0124 00:17:24.987118 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-kw9l8"] Jan 24 00:17:24 crc kubenswrapper[4676]: I0124 00:17:24.988239 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kw9l8" Jan 24 00:17:24 crc kubenswrapper[4676]: I0124 00:17:24.997907 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kw9l8"] Jan 24 00:17:25 crc kubenswrapper[4676]: I0124 00:17:25.004974 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qpx6\" (UniqueName: \"kubernetes.io/projected/25146f99-405a-4473-bf27-69a7195a3338-kube-api-access-6qpx6\") pod \"openstack-operator-index-kw9l8\" (UID: \"25146f99-405a-4473-bf27-69a7195a3338\") " pod="openstack-operators/openstack-operator-index-kw9l8" Jan 24 00:17:25 crc kubenswrapper[4676]: I0124 00:17:25.106462 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qpx6\" (UniqueName: \"kubernetes.io/projected/25146f99-405a-4473-bf27-69a7195a3338-kube-api-access-6qpx6\") pod \"openstack-operator-index-kw9l8\" (UID: \"25146f99-405a-4473-bf27-69a7195a3338\") " pod="openstack-operators/openstack-operator-index-kw9l8" Jan 24 00:17:25 crc kubenswrapper[4676]: I0124 00:17:25.133085 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qpx6\" (UniqueName: \"kubernetes.io/projected/25146f99-405a-4473-bf27-69a7195a3338-kube-api-access-6qpx6\") pod \"openstack-operator-index-kw9l8\" (UID: \"25146f99-405a-4473-bf27-69a7195a3338\") " pod="openstack-operators/openstack-operator-index-kw9l8" Jan 24 00:17:25 crc kubenswrapper[4676]: I0124 00:17:25.317249 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kw9l8" Jan 24 00:17:25 crc kubenswrapper[4676]: I0124 00:17:25.551424 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kw9l8"] Jan 24 00:17:25 crc kubenswrapper[4676]: W0124 00:17:25.557032 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25146f99_405a_4473_bf27_69a7195a3338.slice/crio-17da9ac2c23c5c6cad3d8f4d82d1f4430a5a586f1e9a35c05032d087359c1bd6 WatchSource:0}: Error finding container 17da9ac2c23c5c6cad3d8f4d82d1f4430a5a586f1e9a35c05032d087359c1bd6: Status 404 returned error can't find the container with id 17da9ac2c23c5c6cad3d8f4d82d1f4430a5a586f1e9a35c05032d087359c1bd6 Jan 24 00:17:25 crc kubenswrapper[4676]: I0124 00:17:25.606899 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-s7b59" Jan 24 00:17:25 crc kubenswrapper[4676]: I0124 00:17:25.725761 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-jqk4f" Jan 24 00:17:25 crc kubenswrapper[4676]: I0124 00:17:25.921294 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kw9l8" event={"ID":"25146f99-405a-4473-bf27-69a7195a3338","Type":"ContainerStarted","Data":"17da9ac2c23c5c6cad3d8f4d82d1f4430a5a586f1e9a35c05032d087359c1bd6"} Jan 24 00:17:25 crc kubenswrapper[4676]: I0124 00:17:25.921493 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-6ntrt" podUID="5425e3ea-c929-4943-bd14-d4e193992052" containerName="registry-server" containerID="cri-o://7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d" gracePeriod=2 Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.820581 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6ntrt" Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.852849 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4vfp\" (UniqueName: \"kubernetes.io/projected/5425e3ea-c929-4943-bd14-d4e193992052-kube-api-access-r4vfp\") pod \"5425e3ea-c929-4943-bd14-d4e193992052\" (UID: \"5425e3ea-c929-4943-bd14-d4e193992052\") " Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.861170 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5425e3ea-c929-4943-bd14-d4e193992052-kube-api-access-r4vfp" (OuterVolumeSpecName: "kube-api-access-r4vfp") pod "5425e3ea-c929-4943-bd14-d4e193992052" (UID: "5425e3ea-c929-4943-bd14-d4e193992052"). InnerVolumeSpecName "kube-api-access-r4vfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.930045 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kw9l8" event={"ID":"25146f99-405a-4473-bf27-69a7195a3338","Type":"ContainerStarted","Data":"ae7af5773b0c0c3022a37af8482a6211017babb29f2d4c0ba014e258b1e9ee90"} Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.935305 4676 generic.go:334] "Generic (PLEG): container finished" podID="5425e3ea-c929-4943-bd14-d4e193992052" containerID="7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d" exitCode=0 Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.935354 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6ntrt" event={"ID":"5425e3ea-c929-4943-bd14-d4e193992052","Type":"ContainerDied","Data":"7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d"} Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.935406 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6ntrt" event={"ID":"5425e3ea-c929-4943-bd14-d4e193992052","Type":"ContainerDied","Data":"ee09d2706d4eadeb85bdfce3c04bc857b26be38730265e45634501b3005889aa"} Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.935431 4676 scope.go:117] "RemoveContainer" containerID="7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d" Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.935439 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6ntrt" Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.958495 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4vfp\" (UniqueName: \"kubernetes.io/projected/5425e3ea-c929-4943-bd14-d4e193992052-kube-api-access-r4vfp\") on node \"crc\" DevicePath \"\"" Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.960509 4676 scope.go:117] "RemoveContainer" containerID="7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d" Jan 24 00:17:26 crc kubenswrapper[4676]: E0124 00:17:26.961285 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d\": container with ID starting with 7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d not found: ID does not exist" containerID="7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d" Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.961324 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d"} err="failed to get container status \"7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d\": rpc error: code = NotFound desc = could not find container \"7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d\": container with ID starting with 7921f77247ccf3e14c1bcf7202d66aacc6d4736b70b42e45031192bfc3dc839d not found: ID does not exist" Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.966926 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-kw9l8" podStartSLOduration=2.887173703 podStartE2EDuration="2.966905371s" podCreationTimestamp="2026-01-24 00:17:24 +0000 UTC" firstStartedPulling="2026-01-24 00:17:25.56125536 +0000 UTC m=+829.591226361" lastFinishedPulling="2026-01-24 00:17:25.640987028 +0000 UTC m=+829.670958029" observedRunningTime="2026-01-24 00:17:26.960839375 +0000 UTC m=+830.990810386" watchObservedRunningTime="2026-01-24 00:17:26.966905371 +0000 UTC m=+830.996876392" Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.984206 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6ntrt"] Jan 24 00:17:26 crc kubenswrapper[4676]: I0124 00:17:26.987906 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-6ntrt"] Jan 24 00:17:28 crc kubenswrapper[4676]: I0124 00:17:28.269298 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5425e3ea-c929-4943-bd14-d4e193992052" path="/var/lib/kubelet/pods/5425e3ea-c929-4943-bd14-d4e193992052/volumes" Jan 24 00:17:35 crc kubenswrapper[4676]: I0124 00:17:35.318561 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-kw9l8" Jan 24 00:17:35 crc kubenswrapper[4676]: I0124 00:17:35.319639 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-kw9l8" Jan 24 00:17:35 crc kubenswrapper[4676]: I0124 00:17:35.359546 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-kw9l8" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.057041 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-kw9l8" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.202116 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-szpqt" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.684408 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789"] Jan 24 00:17:36 crc kubenswrapper[4676]: E0124 00:17:36.684680 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5425e3ea-c929-4943-bd14-d4e193992052" containerName="registry-server" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.684700 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="5425e3ea-c929-4943-bd14-d4e193992052" containerName="registry-server" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.684865 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="5425e3ea-c929-4943-bd14-d4e193992052" containerName="registry-server" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.685940 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.690144 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-jhh24" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.710195 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789"] Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.803663 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-util\") pod \"ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.803721 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx2vn\" (UniqueName: \"kubernetes.io/projected/9d669669-26ba-4775-b7cc-e97cc7dbe326-kube-api-access-dx2vn\") pod \"ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.803742 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-bundle\") pod \"ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.905147 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-util\") pod \"ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.905246 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx2vn\" (UniqueName: \"kubernetes.io/projected/9d669669-26ba-4775-b7cc-e97cc7dbe326-kube-api-access-dx2vn\") pod \"ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.905312 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-bundle\") pod \"ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.905765 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-util\") pod \"ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.905782 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-bundle\") pod \"ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:36 crc kubenswrapper[4676]: I0124 00:17:36.925230 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx2vn\" (UniqueName: \"kubernetes.io/projected/9d669669-26ba-4775-b7cc-e97cc7dbe326-kube-api-access-dx2vn\") pod \"ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:37 crc kubenswrapper[4676]: I0124 00:17:37.006054 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:37 crc kubenswrapper[4676]: I0124 00:17:37.444215 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789"] Jan 24 00:17:38 crc kubenswrapper[4676]: I0124 00:17:38.038925 4676 generic.go:334] "Generic (PLEG): container finished" podID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerID="febce8b3323e91f83e00c3dd649ad413f2af7374de871b23981687ad243e2b66" exitCode=0 Jan 24 00:17:38 crc kubenswrapper[4676]: I0124 00:17:38.039111 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" event={"ID":"9d669669-26ba-4775-b7cc-e97cc7dbe326","Type":"ContainerDied","Data":"febce8b3323e91f83e00c3dd649ad413f2af7374de871b23981687ad243e2b66"} Jan 24 00:17:38 crc kubenswrapper[4676]: I0124 00:17:38.039362 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" event={"ID":"9d669669-26ba-4775-b7cc-e97cc7dbe326","Type":"ContainerStarted","Data":"6e080ebc2338efd89d347cecd9a36fade586fddfd61171c9fe2f90594f0f5697"} Jan 24 00:17:39 crc kubenswrapper[4676]: I0124 00:17:39.052081 4676 generic.go:334] "Generic (PLEG): container finished" podID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerID="33cd5040f326fb3e4a05af3f6c86f27b17dad7629e9dd1835e1deaaa7b71a1ab" exitCode=0 Jan 24 00:17:39 crc kubenswrapper[4676]: I0124 00:17:39.052172 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" event={"ID":"9d669669-26ba-4775-b7cc-e97cc7dbe326","Type":"ContainerDied","Data":"33cd5040f326fb3e4a05af3f6c86f27b17dad7629e9dd1835e1deaaa7b71a1ab"} Jan 24 00:17:40 crc kubenswrapper[4676]: I0124 00:17:40.062915 4676 generic.go:334] "Generic (PLEG): container finished" podID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerID="0096dd21fa12fac790c05fdeef3584e1e18ecb50242cb75bfb3734240a194e22" exitCode=0 Jan 24 00:17:40 crc kubenswrapper[4676]: I0124 00:17:40.062967 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" event={"ID":"9d669669-26ba-4775-b7cc-e97cc7dbe326","Type":"ContainerDied","Data":"0096dd21fa12fac790c05fdeef3584e1e18ecb50242cb75bfb3734240a194e22"} Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.351459 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.388820 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx2vn\" (UniqueName: \"kubernetes.io/projected/9d669669-26ba-4775-b7cc-e97cc7dbe326-kube-api-access-dx2vn\") pod \"9d669669-26ba-4775-b7cc-e97cc7dbe326\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.388912 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-util\") pod \"9d669669-26ba-4775-b7cc-e97cc7dbe326\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.388962 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-bundle\") pod \"9d669669-26ba-4775-b7cc-e97cc7dbe326\" (UID: \"9d669669-26ba-4775-b7cc-e97cc7dbe326\") " Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.389950 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-bundle" (OuterVolumeSpecName: "bundle") pod "9d669669-26ba-4775-b7cc-e97cc7dbe326" (UID: "9d669669-26ba-4775-b7cc-e97cc7dbe326"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.395547 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d669669-26ba-4775-b7cc-e97cc7dbe326-kube-api-access-dx2vn" (OuterVolumeSpecName: "kube-api-access-dx2vn") pod "9d669669-26ba-4775-b7cc-e97cc7dbe326" (UID: "9d669669-26ba-4775-b7cc-e97cc7dbe326"). InnerVolumeSpecName "kube-api-access-dx2vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.406740 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-util" (OuterVolumeSpecName: "util") pod "9d669669-26ba-4775-b7cc-e97cc7dbe326" (UID: "9d669669-26ba-4775-b7cc-e97cc7dbe326"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.491305 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx2vn\" (UniqueName: \"kubernetes.io/projected/9d669669-26ba-4775-b7cc-e97cc7dbe326-kube-api-access-dx2vn\") on node \"crc\" DevicePath \"\"" Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.491342 4676 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-util\") on node \"crc\" DevicePath \"\"" Jan 24 00:17:41 crc kubenswrapper[4676]: I0124 00:17:41.491354 4676 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9d669669-26ba-4775-b7cc-e97cc7dbe326-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:17:42 crc kubenswrapper[4676]: I0124 00:17:42.080313 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" event={"ID":"9d669669-26ba-4775-b7cc-e97cc7dbe326","Type":"ContainerDied","Data":"6e080ebc2338efd89d347cecd9a36fade586fddfd61171c9fe2f90594f0f5697"} Jan 24 00:17:42 crc kubenswrapper[4676]: I0124 00:17:42.080677 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e080ebc2338efd89d347cecd9a36fade586fddfd61171c9fe2f90594f0f5697" Jan 24 00:17:42 crc kubenswrapper[4676]: I0124 00:17:42.080456 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.760098 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr"] Jan 24 00:17:44 crc kubenswrapper[4676]: E0124 00:17:44.760822 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerName="pull" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.760841 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerName="pull" Jan 24 00:17:44 crc kubenswrapper[4676]: E0124 00:17:44.760859 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerName="util" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.760869 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerName="util" Jan 24 00:17:44 crc kubenswrapper[4676]: E0124 00:17:44.760895 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerName="extract" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.760905 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerName="extract" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.761059 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d669669-26ba-4775-b7cc-e97cc7dbe326" containerName="extract" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.761633 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.764218 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-w8lw6" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.837085 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjp56\" (UniqueName: \"kubernetes.io/projected/07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6-kube-api-access-fjp56\") pod \"openstack-operator-controller-init-69647cdbc5-96fbr\" (UID: \"07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6\") " pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.861223 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr"] Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.938882 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjp56\" (UniqueName: \"kubernetes.io/projected/07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6-kube-api-access-fjp56\") pod \"openstack-operator-controller-init-69647cdbc5-96fbr\" (UID: \"07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6\") " pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" Jan 24 00:17:44 crc kubenswrapper[4676]: I0124 00:17:44.957208 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjp56\" (UniqueName: \"kubernetes.io/projected/07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6-kube-api-access-fjp56\") pod \"openstack-operator-controller-init-69647cdbc5-96fbr\" (UID: \"07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6\") " pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" Jan 24 00:17:45 crc kubenswrapper[4676]: I0124 00:17:45.076682 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" Jan 24 00:17:45 crc kubenswrapper[4676]: I0124 00:17:45.555195 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr"] Jan 24 00:17:46 crc kubenswrapper[4676]: I0124 00:17:46.110492 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" event={"ID":"07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6","Type":"ContainerStarted","Data":"d5895aa75d2cdcc2aa52ae1b3ab7d21be12e0afd307d310c4434c4bef4953f6c"} Jan 24 00:17:50 crc kubenswrapper[4676]: I0124 00:17:50.146367 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" event={"ID":"07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6","Type":"ContainerStarted","Data":"7cf047d5cb51ded3ea0d506ac23e1396d51d95d1b31f3ab0ba6365be60ad4afd"} Jan 24 00:17:50 crc kubenswrapper[4676]: I0124 00:17:50.146919 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" Jan 24 00:17:50 crc kubenswrapper[4676]: I0124 00:17:50.178351 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" podStartSLOduration=1.752634459 podStartE2EDuration="6.178334955s" podCreationTimestamp="2026-01-24 00:17:44 +0000 UTC" firstStartedPulling="2026-01-24 00:17:45.562475097 +0000 UTC m=+849.592446108" lastFinishedPulling="2026-01-24 00:17:49.988175603 +0000 UTC m=+854.018146604" observedRunningTime="2026-01-24 00:17:50.173692392 +0000 UTC m=+854.203663403" watchObservedRunningTime="2026-01-24 00:17:50.178334955 +0000 UTC m=+854.208305966" Jan 24 00:17:55 crc kubenswrapper[4676]: I0124 00:17:55.080135 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-69647cdbc5-96fbr" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.234554 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.236051 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.238021 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-f2s6k" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.246619 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.266098 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.266728 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.276592 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-svdrx" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.290049 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.303824 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.320151 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.337389 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-g8q5n" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.343046 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.343788 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.345900 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-ncsvx" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.346803 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsns8\" (UniqueName: \"kubernetes.io/projected/02123851-7d2f-477b-9c60-5a9922a0bc97-kube-api-access-bsns8\") pod \"barbican-operator-controller-manager-7f86f8796f-fgplq\" (UID: \"02123851-7d2f-477b-9c60-5a9922a0bc97\") " pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.346921 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92kpk\" (UniqueName: \"kubernetes.io/projected/29e4b64d-19bd-419b-9e29-7a41e6f12ae0-kube-api-access-92kpk\") pod \"glance-operator-controller-manager-78fdd796fd-btxnv\" (UID: \"29e4b64d-19bd-419b-9e29-7a41e6f12ae0\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.347005 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7g5p\" (UniqueName: \"kubernetes.io/projected/0cd05b9f-6699-46e3-ae36-9f21352e6c8e-kube-api-access-q7g5p\") pod \"cinder-operator-controller-manager-69cf5d4557-c8c6m\" (UID: \"0cd05b9f-6699-46e3-ae36-9f21352e6c8e\") " pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.373219 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.384948 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.386111 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.391613 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.409272 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.415046 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-r5w45" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.426890 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.427672 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.430847 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.431481 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.448603 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f9hs\" (UniqueName: \"kubernetes.io/projected/6b8541f9-a37a-41d6-8006-3d0335c3abb5-kube-api-access-6f9hs\") pod \"horizon-operator-controller-manager-77d5c5b54f-rbzsj\" (UID: \"6b8541f9-a37a-41d6-8006-3d0335c3abb5\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.448649 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw5zz\" (UniqueName: \"kubernetes.io/projected/921e121c-5261-4fe7-8171-6b634babedf4-kube-api-access-hw5zz\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.448694 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsns8\" (UniqueName: \"kubernetes.io/projected/02123851-7d2f-477b-9c60-5a9922a0bc97-kube-api-access-bsns8\") pod \"barbican-operator-controller-manager-7f86f8796f-fgplq\" (UID: \"02123851-7d2f-477b-9c60-5a9922a0bc97\") " pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.448721 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.448743 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsl2n\" (UniqueName: \"kubernetes.io/projected/5e9cf1cb-c413-45ad-8a51-bf35407fcdfe-kube-api-access-qsl2n\") pod \"heat-operator-controller-manager-594c8c9d5d-p2nr8\" (UID: \"5e9cf1cb-c413-45ad-8a51-bf35407fcdfe\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.448771 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2lp4\" (UniqueName: \"kubernetes.io/projected/a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85-kube-api-access-c2lp4\") pod \"designate-operator-controller-manager-b45d7bf98-js6db\" (UID: \"a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.448795 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92kpk\" (UniqueName: \"kubernetes.io/projected/29e4b64d-19bd-419b-9e29-7a41e6f12ae0-kube-api-access-92kpk\") pod \"glance-operator-controller-manager-78fdd796fd-btxnv\" (UID: \"29e4b64d-19bd-419b-9e29-7a41e6f12ae0\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.448822 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7g5p\" (UniqueName: \"kubernetes.io/projected/0cd05b9f-6699-46e3-ae36-9f21352e6c8e-kube-api-access-q7g5p\") pod \"cinder-operator-controller-manager-69cf5d4557-c8c6m\" (UID: \"0cd05b9f-6699-46e3-ae36-9f21352e6c8e\") " pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.456268 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.470071 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.473890 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.474414 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kcp2q" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.474658 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.474871 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-j96cn" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.480718 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-69mq8" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.493424 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.511258 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.515060 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7g5p\" (UniqueName: \"kubernetes.io/projected/0cd05b9f-6699-46e3-ae36-9f21352e6c8e-kube-api-access-q7g5p\") pod \"cinder-operator-controller-manager-69cf5d4557-c8c6m\" (UID: \"0cd05b9f-6699-46e3-ae36-9f21352e6c8e\") " pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.515650 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsns8\" (UniqueName: \"kubernetes.io/projected/02123851-7d2f-477b-9c60-5a9922a0bc97-kube-api-access-bsns8\") pod \"barbican-operator-controller-manager-7f86f8796f-fgplq\" (UID: \"02123851-7d2f-477b-9c60-5a9922a0bc97\") " pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.524781 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.526639 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.544967 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-g9dr2" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.551444 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.552152 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2lp4\" (UniqueName: \"kubernetes.io/projected/a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85-kube-api-access-c2lp4\") pod \"designate-operator-controller-manager-b45d7bf98-js6db\" (UID: \"a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.552223 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f9hs\" (UniqueName: \"kubernetes.io/projected/6b8541f9-a37a-41d6-8006-3d0335c3abb5-kube-api-access-6f9hs\") pod \"horizon-operator-controller-manager-77d5c5b54f-rbzsj\" (UID: \"6b8541f9-a37a-41d6-8006-3d0335c3abb5\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.552247 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fqcx\" (UniqueName: \"kubernetes.io/projected/555ebb8f-1bc3-4b8d-9f37-cad92b48477c-kube-api-access-5fqcx\") pod \"ironic-operator-controller-manager-598f7747c9-v25h4\" (UID: \"555ebb8f-1bc3-4b8d-9f37-cad92b48477c\") " pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.552285 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw5zz\" (UniqueName: \"kubernetes.io/projected/921e121c-5261-4fe7-8171-6b634babedf4-kube-api-access-hw5zz\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.552314 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np4zv\" (UniqueName: \"kubernetes.io/projected/9a3f9a14-1138-425d-8a56-454b282d7d9f-kube-api-access-np4zv\") pod \"keystone-operator-controller-manager-b8b6d4659-2gzqg\" (UID: \"9a3f9a14-1138-425d-8a56-454b282d7d9f\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.552360 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.552410 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsl2n\" (UniqueName: \"kubernetes.io/projected/5e9cf1cb-c413-45ad-8a51-bf35407fcdfe-kube-api-access-qsl2n\") pod \"heat-operator-controller-manager-594c8c9d5d-p2nr8\" (UID: \"5e9cf1cb-c413-45ad-8a51-bf35407fcdfe\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" Jan 24 00:18:16 crc kubenswrapper[4676]: E0124 00:18:16.552593 4676 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:16 crc kubenswrapper[4676]: E0124 00:18:16.552667 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert podName:921e121c-5261-4fe7-8171-6b634babedf4 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:17.052647037 +0000 UTC m=+881.082618038 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert") pod "infra-operator-controller-manager-58749ffdfb-jxx26" (UID: "921e121c-5261-4fe7-8171-6b634babedf4") : secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.565714 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.583102 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92kpk\" (UniqueName: \"kubernetes.io/projected/29e4b64d-19bd-419b-9e29-7a41e6f12ae0-kube-api-access-92kpk\") pod \"glance-operator-controller-manager-78fdd796fd-btxnv\" (UID: \"29e4b64d-19bd-419b-9e29-7a41e6f12ae0\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.585980 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsl2n\" (UniqueName: \"kubernetes.io/projected/5e9cf1cb-c413-45ad-8a51-bf35407fcdfe-kube-api-access-qsl2n\") pod \"heat-operator-controller-manager-594c8c9d5d-p2nr8\" (UID: \"5e9cf1cb-c413-45ad-8a51-bf35407fcdfe\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.589872 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.590566 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.598727 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f9hs\" (UniqueName: \"kubernetes.io/projected/6b8541f9-a37a-41d6-8006-3d0335c3abb5-kube-api-access-6f9hs\") pod \"horizon-operator-controller-manager-77d5c5b54f-rbzsj\" (UID: \"6b8541f9-a37a-41d6-8006-3d0335c3abb5\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.601544 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.606504 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.608961 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw5zz\" (UniqueName: \"kubernetes.io/projected/921e121c-5261-4fe7-8171-6b634babedf4-kube-api-access-hw5zz\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.619144 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2lp4\" (UniqueName: \"kubernetes.io/projected/a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85-kube-api-access-c2lp4\") pod \"designate-operator-controller-manager-b45d7bf98-js6db\" (UID: \"a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.619901 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-r2mlc" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.653101 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fqcx\" (UniqueName: \"kubernetes.io/projected/555ebb8f-1bc3-4b8d-9f37-cad92b48477c-kube-api-access-5fqcx\") pod \"ironic-operator-controller-manager-598f7747c9-v25h4\" (UID: \"555ebb8f-1bc3-4b8d-9f37-cad92b48477c\") " pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.653165 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np4zv\" (UniqueName: \"kubernetes.io/projected/9a3f9a14-1138-425d-8a56-454b282d7d9f-kube-api-access-np4zv\") pod \"keystone-operator-controller-manager-b8b6d4659-2gzqg\" (UID: \"9a3f9a14-1138-425d-8a56-454b282d7d9f\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.653242 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8zz7\" (UniqueName: \"kubernetes.io/projected/dfc79179-d245-4360-be6e-8b43441e23ed-kube-api-access-l8zz7\") pod \"manila-operator-controller-manager-78c6999f6f-g4d8z\" (UID: \"dfc79179-d245-4360-be6e-8b43441e23ed\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.661630 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.667680 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.678160 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.679197 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.696887 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-78hg7" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.700792 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.706140 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.706636 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np4zv\" (UniqueName: \"kubernetes.io/projected/9a3f9a14-1138-425d-8a56-454b282d7d9f-kube-api-access-np4zv\") pod \"keystone-operator-controller-manager-b8b6d4659-2gzqg\" (UID: \"9a3f9a14-1138-425d-8a56-454b282d7d9f\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.707226 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.708908 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.718058 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qc7mt" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.718297 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fqcx\" (UniqueName: \"kubernetes.io/projected/555ebb8f-1bc3-4b8d-9f37-cad92b48477c-kube-api-access-5fqcx\") pod \"ironic-operator-controller-manager-598f7747c9-v25h4\" (UID: \"555ebb8f-1bc3-4b8d-9f37-cad92b48477c\") " pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.725843 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.726634 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.734783 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-gqgzt" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.735517 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.753885 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.765063 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvgz4\" (UniqueName: \"kubernetes.io/projected/4ce661f6-26e2-4da2-a759-e493a60587b2-kube-api-access-xvgz4\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n\" (UID: \"4ce661f6-26e2-4da2-a759-e493a60587b2\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.765118 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8zz7\" (UniqueName: \"kubernetes.io/projected/dfc79179-d245-4360-be6e-8b43441e23ed-kube-api-access-l8zz7\") pod \"manila-operator-controller-manager-78c6999f6f-g4d8z\" (UID: \"dfc79179-d245-4360-be6e-8b43441e23ed\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.765198 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8pmb\" (UniqueName: \"kubernetes.io/projected/df9ab5f0-f577-4303-8045-f960c67a6936-kube-api-access-l8pmb\") pod \"nova-operator-controller-manager-6b8bc8d87d-4xp45\" (UID: \"df9ab5f0-f577-4303-8045-f960c67a6936\") " pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.765263 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22p7t\" (UniqueName: \"kubernetes.io/projected/dd6346d8-9cf1-4364-b480-f4c2d872472f-kube-api-access-22p7t\") pod \"neutron-operator-controller-manager-78d58447c5-wqbcz\" (UID: \"dd6346d8-9cf1-4364-b480-f4c2d872472f\") " pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.767546 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.771647 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.774360 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.775943 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-b62h5" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.792904 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.844560 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8zz7\" (UniqueName: \"kubernetes.io/projected/dfc79179-d245-4360-be6e-8b43441e23ed-kube-api-access-l8zz7\") pod \"manila-operator-controller-manager-78c6999f6f-g4d8z\" (UID: \"dfc79179-d245-4360-be6e-8b43441e23ed\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.874059 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22p7t\" (UniqueName: \"kubernetes.io/projected/dd6346d8-9cf1-4364-b480-f4c2d872472f-kube-api-access-22p7t\") pod \"neutron-operator-controller-manager-78d58447c5-wqbcz\" (UID: \"dd6346d8-9cf1-4364-b480-f4c2d872472f\") " pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.874102 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvgz4\" (UniqueName: \"kubernetes.io/projected/4ce661f6-26e2-4da2-a759-e493a60587b2-kube-api-access-xvgz4\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n\" (UID: \"4ce661f6-26e2-4da2-a759-e493a60587b2\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.874161 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8pmb\" (UniqueName: \"kubernetes.io/projected/df9ab5f0-f577-4303-8045-f960c67a6936-kube-api-access-l8pmb\") pod \"nova-operator-controller-manager-6b8bc8d87d-4xp45\" (UID: \"df9ab5f0-f577-4303-8045-f960c67a6936\") " pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.874192 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5lfg\" (UniqueName: \"kubernetes.io/projected/ced74bcb-8345-40c5-b2d4-3d369f30b835-kube-api-access-c5lfg\") pod \"octavia-operator-controller-manager-7bd9774b6-k8lw7\" (UID: \"ced74bcb-8345-40c5-b2d4-3d369f30b835\") " pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.875610 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.876295 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.891226 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-dpjc9" Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.932087 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms"] Jan 24 00:18:16 crc kubenswrapper[4676]: I0124 00:18:16.953142 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.014183 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5lfg\" (UniqueName: \"kubernetes.io/projected/ced74bcb-8345-40c5-b2d4-3d369f30b835-kube-api-access-c5lfg\") pod \"octavia-operator-controller-manager-7bd9774b6-k8lw7\" (UID: \"ced74bcb-8345-40c5-b2d4-3d369f30b835\") " pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.026505 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.026938 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rtqkf" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.026596 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.036205 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.036539 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.037058 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22p7t\" (UniqueName: \"kubernetes.io/projected/dd6346d8-9cf1-4364-b480-f4c2d872472f-kube-api-access-22p7t\") pod \"neutron-operator-controller-manager-78d58447c5-wqbcz\" (UID: \"dd6346d8-9cf1-4364-b480-f4c2d872472f\") " pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.037899 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8pmb\" (UniqueName: \"kubernetes.io/projected/df9ab5f0-f577-4303-8045-f960c67a6936-kube-api-access-l8pmb\") pod \"nova-operator-controller-manager-6b8bc8d87d-4xp45\" (UID: \"df9ab5f0-f577-4303-8045-f960c67a6936\") " pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.047896 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvgz4\" (UniqueName: \"kubernetes.io/projected/4ce661f6-26e2-4da2-a759-e493a60587b2-kube-api-access-xvgz4\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n\" (UID: \"4ce661f6-26e2-4da2-a759-e493a60587b2\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.061268 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.061801 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqqxj\" (UniqueName: \"kubernetes.io/projected/53358678-d763-4b02-a157-86a57ebd0305-kube-api-access-wqqxj\") pod \"ovn-operator-controller-manager-55db956ddc-49nrq\" (UID: \"53358678-d763-4b02-a157-86a57ebd0305\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.077612 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.082576 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.087968 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.105753 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.107720 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-kmhkx" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.136835 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.153492 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.166019 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pggf\" (UniqueName: \"kubernetes.io/projected/d85fa79d-818f-4079-aac4-f3fa51a90e9a-kube-api-access-4pggf\") pod \"placement-operator-controller-manager-5d646b7d76-tdqwh\" (UID: \"d85fa79d-818f-4079-aac4-f3fa51a90e9a\") " pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.166060 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.166119 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.166141 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgls7\" (UniqueName: \"kubernetes.io/projected/196f45b9-e656-4760-b058-e0b5c08a50d9-kube-api-access-jgls7\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.166183 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqqxj\" (UniqueName: \"kubernetes.io/projected/53358678-d763-4b02-a157-86a57ebd0305-kube-api-access-wqqxj\") pod \"ovn-operator-controller-manager-55db956ddc-49nrq\" (UID: \"53358678-d763-4b02-a157-86a57ebd0305\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.166500 4676 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.166556 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert podName:921e121c-5261-4fe7-8171-6b634babedf4 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:18.166540183 +0000 UTC m=+882.196511184 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert") pod "infra-operator-controller-manager-58749ffdfb-jxx26" (UID: "921e121c-5261-4fe7-8171-6b634babedf4") : secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.172308 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5lfg\" (UniqueName: \"kubernetes.io/projected/ced74bcb-8345-40c5-b2d4-3d369f30b835-kube-api-access-c5lfg\") pod \"octavia-operator-controller-manager-7bd9774b6-k8lw7\" (UID: \"ced74bcb-8345-40c5-b2d4-3d369f30b835\") " pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.175408 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.198265 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqqxj\" (UniqueName: \"kubernetes.io/projected/53358678-d763-4b02-a157-86a57ebd0305-kube-api-access-wqqxj\") pod \"ovn-operator-controller-manager-55db956ddc-49nrq\" (UID: \"53358678-d763-4b02-a157-86a57ebd0305\") " pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.199927 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.200733 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.211097 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-xd5bv" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.242813 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.267474 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j5xn\" (UniqueName: \"kubernetes.io/projected/b0c8972b-31d7-40c1-bc65-1478718d41a5-kube-api-access-8j5xn\") pod \"swift-operator-controller-manager-5df95d5965-h8wx9\" (UID: \"b0c8972b-31d7-40c1-bc65-1478718d41a5\") " pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.267531 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pggf\" (UniqueName: \"kubernetes.io/projected/d85fa79d-818f-4079-aac4-f3fa51a90e9a-kube-api-access-4pggf\") pod \"placement-operator-controller-manager-5d646b7d76-tdqwh\" (UID: \"d85fa79d-818f-4079-aac4-f3fa51a90e9a\") " pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.267565 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.267646 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgls7\" (UniqueName: \"kubernetes.io/projected/196f45b9-e656-4760-b058-e0b5c08a50d9-kube-api-access-jgls7\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.268056 4676 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.268097 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert podName:196f45b9-e656-4760-b058-e0b5c08a50d9 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:17.768083302 +0000 UTC m=+881.798054303 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" (UID: "196f45b9-e656-4760-b058-e0b5c08a50d9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.291525 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.292329 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.293642 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.321873 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-v4559" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.343894 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgls7\" (UniqueName: \"kubernetes.io/projected/196f45b9-e656-4760-b058-e0b5c08a50d9-kube-api-access-jgls7\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.364432 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.365276 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.373297 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j5xn\" (UniqueName: \"kubernetes.io/projected/b0c8972b-31d7-40c1-bc65-1478718d41a5-kube-api-access-8j5xn\") pod \"swift-operator-controller-manager-5df95d5965-h8wx9\" (UID: \"b0c8972b-31d7-40c1-bc65-1478718d41a5\") " pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.373405 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvlxx\" (UniqueName: \"kubernetes.io/projected/060e1c8d-dfa6-428f-bffe-d89ac3dab8c3-kube-api-access-qvlxx\") pod \"telemetry-operator-controller-manager-85cd9769bb-4c5zl\" (UID: \"060e1c8d-dfa6-428f-bffe-d89ac3dab8c3\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.374944 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-hmp8k" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.386991 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pggf\" (UniqueName: \"kubernetes.io/projected/d85fa79d-818f-4079-aac4-f3fa51a90e9a-kube-api-access-4pggf\") pod \"placement-operator-controller-manager-5d646b7d76-tdqwh\" (UID: \"d85fa79d-818f-4079-aac4-f3fa51a90e9a\") " pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.397008 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.430373 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.431300 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.439149 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-jkv69" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.441438 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.443029 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.444617 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j5xn\" (UniqueName: \"kubernetes.io/projected/b0c8972b-31d7-40c1-bc65-1478718d41a5-kube-api-access-8j5xn\") pod \"swift-operator-controller-manager-5df95d5965-h8wx9\" (UID: \"b0c8972b-31d7-40c1-bc65-1478718d41a5\") " pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.475864 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvlxx\" (UniqueName: \"kubernetes.io/projected/060e1c8d-dfa6-428f-bffe-d89ac3dab8c3-kube-api-access-qvlxx\") pod \"telemetry-operator-controller-manager-85cd9769bb-4c5zl\" (UID: \"060e1c8d-dfa6-428f-bffe-d89ac3dab8c3\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.476045 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xd87\" (UniqueName: \"kubernetes.io/projected/6245b73e-9fba-4ad7-bbbc-31db48c03825-kube-api-access-2xd87\") pod \"test-operator-controller-manager-69797bbcbd-gqg82\" (UID: \"6245b73e-9fba-4ad7-bbbc-31db48c03825\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.481446 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.499958 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.500797 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.501729 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvlxx\" (UniqueName: \"kubernetes.io/projected/060e1c8d-dfa6-428f-bffe-d89ac3dab8c3-kube-api-access-qvlxx\") pod \"telemetry-operator-controller-manager-85cd9769bb-4c5zl\" (UID: \"060e1c8d-dfa6-428f-bffe-d89ac3dab8c3\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.510881 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.511046 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-2p4mv" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.511186 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.546223 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.576992 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.577041 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xd87\" (UniqueName: \"kubernetes.io/projected/6245b73e-9fba-4ad7-bbbc-31db48c03825-kube-api-access-2xd87\") pod \"test-operator-controller-manager-69797bbcbd-gqg82\" (UID: \"6245b73e-9fba-4ad7-bbbc-31db48c03825\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.577090 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjbfn\" (UniqueName: \"kubernetes.io/projected/9d79c791-c851-4c4a-aa2d-d175b668b0f5-kube-api-access-cjbfn\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.577122 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmsrh\" (UniqueName: \"kubernetes.io/projected/ccb1ff12-bef7-4f23-b084-fae32f8202ac-kube-api-access-hmsrh\") pod \"watcher-operator-controller-manager-6d9458688d-h6hzt\" (UID: \"ccb1ff12-bef7-4f23-b084-fae32f8202ac\") " pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.577152 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.582367 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.596079 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xd87\" (UniqueName: \"kubernetes.io/projected/6245b73e-9fba-4ad7-bbbc-31db48c03825-kube-api-access-2xd87\") pod \"test-operator-controller-manager-69797bbcbd-gqg82\" (UID: \"6245b73e-9fba-4ad7-bbbc-31db48c03825\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.624496 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.625552 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.626577 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.644802 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-pplpv" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.682743 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.693805 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.693888 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjbfn\" (UniqueName: \"kubernetes.io/projected/9d79c791-c851-4c4a-aa2d-d175b668b0f5-kube-api-access-cjbfn\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.693938 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmsrh\" (UniqueName: \"kubernetes.io/projected/ccb1ff12-bef7-4f23-b084-fae32f8202ac-kube-api-access-hmsrh\") pod \"watcher-operator-controller-manager-6d9458688d-h6hzt\" (UID: \"ccb1ff12-bef7-4f23-b084-fae32f8202ac\") " pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.693995 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.694031 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ffff\" (UniqueName: \"kubernetes.io/projected/fce4c8d0-b903-4873-8c89-2f4b9dd9c05d-kube-api-access-8ffff\") pod \"rabbitmq-cluster-operator-manager-668c99d594-h44qw\" (UID: \"fce4c8d0-b903-4873-8c89-2f4b9dd9c05d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.694195 4676 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.694240 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:18.1942269 +0000 UTC m=+882.224197901 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "metrics-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.694719 4676 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.694749 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:18.194742586 +0000 UTC m=+882.224713587 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "webhook-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.697704 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.726127 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.732093 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmsrh\" (UniqueName: \"kubernetes.io/projected/ccb1ff12-bef7-4f23-b084-fae32f8202ac-kube-api-access-hmsrh\") pod \"watcher-operator-controller-manager-6d9458688d-h6hzt\" (UID: \"ccb1ff12-bef7-4f23-b084-fae32f8202ac\") " pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.742549 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjbfn\" (UniqueName: \"kubernetes.io/projected/9d79c791-c851-4c4a-aa2d-d175b668b0f5-kube-api-access-cjbfn\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.761519 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.796656 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.796748 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ffff\" (UniqueName: \"kubernetes.io/projected/fce4c8d0-b903-4873-8c89-2f4b9dd9c05d-kube-api-access-8ffff\") pod \"rabbitmq-cluster-operator-manager-668c99d594-h44qw\" (UID: \"fce4c8d0-b903-4873-8c89-2f4b9dd9c05d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.797064 4676 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: E0124 00:18:17.797100 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert podName:196f45b9-e656-4760-b058-e0b5c08a50d9 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:18.79708783 +0000 UTC m=+882.827058831 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" (UID: "196f45b9-e656-4760-b058-e0b5c08a50d9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.854106 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ffff\" (UniqueName: \"kubernetes.io/projected/fce4c8d0-b903-4873-8c89-2f4b9dd9c05d-kube-api-access-8ffff\") pod \"rabbitmq-cluster-operator-manager-668c99d594-h44qw\" (UID: \"fce4c8d0-b903-4873-8c89-2f4b9dd9c05d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.876935 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m"] Jan 24 00:18:17 crc kubenswrapper[4676]: I0124 00:18:17.993778 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.105000 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.203412 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.209572 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.209653 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.209681 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.209809 4676 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.209862 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:19.209848665 +0000 UTC m=+883.239819666 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "webhook-server-cert" not found Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.210156 4676 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.210180 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:19.210173266 +0000 UTC m=+883.240144257 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "metrics-server-cert" not found Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.210213 4676 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.210230 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert podName:921e121c-5261-4fe7-8171-6b634babedf4 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:20.210224347 +0000 UTC m=+884.240195348 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert") pod "infra-operator-controller-manager-58749ffdfb-jxx26" (UID: "921e121c-5261-4fe7-8171-6b634babedf4") : secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.272020 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.384401 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" event={"ID":"02123851-7d2f-477b-9c60-5a9922a0bc97","Type":"ContainerStarted","Data":"e20c527f30f5d5b32ca5949019111e8802f9032a5b2f39192183a1b0f159bcc5"} Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.385173 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" event={"ID":"29e4b64d-19bd-419b-9e29-7a41e6f12ae0","Type":"ContainerStarted","Data":"44e586eb58f980cd9a0ede656a3c8c33e098e8bc63066dbec3a5a469f4584367"} Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.385911 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" event={"ID":"0cd05b9f-6699-46e3-ae36-9f21352e6c8e","Type":"ContainerStarted","Data":"45c0fe6a9ceab2d1ac73fa5eaae6830d04b6cd08af0aa6c202dcd877e271a46c"} Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.386753 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" event={"ID":"6b8541f9-a37a-41d6-8006-3d0335c3abb5","Type":"ContainerStarted","Data":"45eaa4d43cf04521f914a5f95aeda6ec7af6d7e2b2dc5d614004f85e53d73860"} Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.537848 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.574276 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.584074 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.597720 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.604741 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.608576 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.614454 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.617994 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7"] Jan 24 00:18:18 crc kubenswrapper[4676]: W0124 00:18:18.624668 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9f1e2a4_c9fa_4136_aa76_059dc2ed9c85.slice/crio-491d48398e5cb5edcfd79ff249314d3e12da8e20e4a48de40e4ece0c05936d99 WatchSource:0}: Error finding container 491d48398e5cb5edcfd79ff249314d3e12da8e20e4a48de40e4ece0c05936d99: Status 404 returned error can't find the container with id 491d48398e5cb5edcfd79ff249314d3e12da8e20e4a48de40e4ece0c05936d99 Jan 24 00:18:18 crc kubenswrapper[4676]: W0124 00:18:18.643561 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddfc79179_d245_4360_be6e_8b43441e23ed.slice/crio-1effa79b6d68858cf3890e790e598a1b0b49afa63e8a1e6465db497b22d7dcef WatchSource:0}: Error finding container 1effa79b6d68858cf3890e790e598a1b0b49afa63e8a1e6465db497b22d7dcef: Status 404 returned error can't find the container with id 1effa79b6d68858cf3890e790e598a1b0b49afa63e8a1e6465db497b22d7dcef Jan 24 00:18:18 crc kubenswrapper[4676]: W0124 00:18:18.650974 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podced74bcb_8345_40c5_b2d4_3d369f30b835.slice/crio-8c0846728bf7c6460610200c6b0a93cf611aa732e7e5e69fb5b0bd2528c44609 WatchSource:0}: Error finding container 8c0846728bf7c6460610200c6b0a93cf611aa732e7e5e69fb5b0bd2528c44609: Status 404 returned error can't find the container with id 8c0846728bf7c6460610200c6b0a93cf611aa732e7e5e69fb5b0bd2528c44609 Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.749391 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.752589 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9"] Jan 24 00:18:18 crc kubenswrapper[4676]: W0124 00:18:18.768820 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0c8972b_31d7_40c1_bc65_1478718d41a5.slice/crio-a078b2e778dd350d037e4f19a131ef8e94c2214c9bdcaa32c53c9fc68b0e9bcd WatchSource:0}: Error finding container a078b2e778dd350d037e4f19a131ef8e94c2214c9bdcaa32c53c9fc68b0e9bcd: Status 404 returned error can't find the container with id a078b2e778dd350d037e4f19a131ef8e94c2214c9bdcaa32c53c9fc68b0e9bcd Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.795467 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45"] Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.800860 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz"] Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.803787 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:4e995cfa360a9d595a01b9c0541ab934692f2374203cb5738127dd784f793831,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l8pmb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-6b8bc8d87d-4xp45_openstack-operators(df9ab5f0-f577-4303-8045-f960c67a6936): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.807594 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" podUID="df9ab5f0-f577-4303-8045-f960c67a6936" Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.823477 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.823664 4676 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.823728 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert podName:196f45b9-e656-4760-b058-e0b5c08a50d9 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:20.82370877 +0000 UTC m=+884.853679771 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" (UID: "196f45b9-e656-4760-b058-e0b5c08a50d9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.833553 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:816d474f502d730d6a2522a272b0e09a2d579ac63617817655d60c54bda4191e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-22p7t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-78d58447c5-wqbcz_openstack-operators(dd6346d8-9cf1-4364-b480-f4c2d872472f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.836534 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" podUID="dd6346d8-9cf1-4364-b480-f4c2d872472f" Jan 24 00:18:18 crc kubenswrapper[4676]: I0124 00:18:18.844126 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh"] Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.853308 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4pggf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5d646b7d76-tdqwh_openstack-operators(d85fa79d-818f-4079-aac4-f3fa51a90e9a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 24 00:18:18 crc kubenswrapper[4676]: E0124 00:18:18.856306 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" podUID="d85fa79d-818f-4079-aac4-f3fa51a90e9a" Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.066739 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt"] Jan 24 00:18:19 crc kubenswrapper[4676]: W0124 00:18:19.073704 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccb1ff12_bef7_4f23_b084_fae32f8202ac.slice/crio-b1531e8c152f4f320cfbdfbb33ee66a0f7d7f32bae2533101682bcec0d01b59a WatchSource:0}: Error finding container b1531e8c152f4f320cfbdfbb33ee66a0f7d7f32bae2533101682bcec0d01b59a: Status 404 returned error can't find the container with id b1531e8c152f4f320cfbdfbb33ee66a0f7d7f32bae2533101682bcec0d01b59a Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.080487 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw"] Jan 24 00:18:19 crc kubenswrapper[4676]: W0124 00:18:19.087756 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfce4c8d0_b903_4873_8c89_2f4b9dd9c05d.slice/crio-4eff56239fa4d23e04006207267c571b42ebc5d352c2fd76046ae95a353f124c WatchSource:0}: Error finding container 4eff56239fa4d23e04006207267c571b42ebc5d352c2fd76046ae95a353f124c: Status 404 returned error can't find the container with id 4eff56239fa4d23e04006207267c571b42ebc5d352c2fd76046ae95a353f124c Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.096144 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82"] Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.099008 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8ffff,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-h44qw_openstack-operators(fce4c8d0-b903-4873-8c89-2f4b9dd9c05d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.100130 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" podUID="fce4c8d0-b903-4873-8c89-2f4b9dd9c05d" Jan 24 00:18:19 crc kubenswrapper[4676]: W0124 00:18:19.100483 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6245b73e_9fba_4ad7_bbbc_31db48c03825.slice/crio-7731d3601c99b3906aa086320f86fb3b157d3282fd194cb0c802ec8777cf95a8 WatchSource:0}: Error finding container 7731d3601c99b3906aa086320f86fb3b157d3282fd194cb0c802ec8777cf95a8: Status 404 returned error can't find the container with id 7731d3601c99b3906aa086320f86fb3b157d3282fd194cb0c802ec8777cf95a8 Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.102282 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2xd87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-69797bbcbd-gqg82_openstack-operators(6245b73e-9fba-4ad7-bbbc-31db48c03825): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.103610 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" podUID="6245b73e-9fba-4ad7-bbbc-31db48c03825" Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.229238 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.229319 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.229440 4676 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.229454 4676 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.229513 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:21.22950014 +0000 UTC m=+885.259471131 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "metrics-server-cert" not found Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.229528 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:21.22952337 +0000 UTC m=+885.259494371 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "webhook-server-cert" not found Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.403369 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" event={"ID":"fce4c8d0-b903-4873-8c89-2f4b9dd9c05d","Type":"ContainerStarted","Data":"4eff56239fa4d23e04006207267c571b42ebc5d352c2fd76046ae95a353f124c"} Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.406073 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" podUID="fce4c8d0-b903-4873-8c89-2f4b9dd9c05d" Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.407560 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" event={"ID":"ced74bcb-8345-40c5-b2d4-3d369f30b835","Type":"ContainerStarted","Data":"8c0846728bf7c6460610200c6b0a93cf611aa732e7e5e69fb5b0bd2528c44609"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.411348 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" event={"ID":"4ce661f6-26e2-4da2-a759-e493a60587b2","Type":"ContainerStarted","Data":"f1039526aec7f470820993e29e55c188bc4ec94d2c8293216741654459984a74"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.422973 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" event={"ID":"dfc79179-d245-4360-be6e-8b43441e23ed","Type":"ContainerStarted","Data":"1effa79b6d68858cf3890e790e598a1b0b49afa63e8a1e6465db497b22d7dcef"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.433497 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" event={"ID":"b0c8972b-31d7-40c1-bc65-1478718d41a5","Type":"ContainerStarted","Data":"a078b2e778dd350d037e4f19a131ef8e94c2214c9bdcaa32c53c9fc68b0e9bcd"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.445651 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" event={"ID":"ccb1ff12-bef7-4f23-b084-fae32f8202ac","Type":"ContainerStarted","Data":"b1531e8c152f4f320cfbdfbb33ee66a0f7d7f32bae2533101682bcec0d01b59a"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.446621 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" event={"ID":"d85fa79d-818f-4079-aac4-f3fa51a90e9a","Type":"ContainerStarted","Data":"4d82172944031bfd3d897c7e4534f048812b4aa5b3bc8fda4f12eaeb9d62c1ff"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.448632 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" event={"ID":"a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85","Type":"ContainerStarted","Data":"491d48398e5cb5edcfd79ff249314d3e12da8e20e4a48de40e4ece0c05936d99"} Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.451783 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" podUID="d85fa79d-818f-4079-aac4-f3fa51a90e9a" Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.452459 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" event={"ID":"df9ab5f0-f577-4303-8045-f960c67a6936","Type":"ContainerStarted","Data":"3ce58638fcbb3d58412b193afdac8082c6db1cc56b208770ffcde92f9729f9eb"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.453432 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" event={"ID":"5e9cf1cb-c413-45ad-8a51-bf35407fcdfe","Type":"ContainerStarted","Data":"11c933af0ebf415f4a79c80ce49d3a9fc82063b843f7764cff971c55914a754b"} Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.456362 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:4e995cfa360a9d595a01b9c0541ab934692f2374203cb5738127dd784f793831\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" podUID="df9ab5f0-f577-4303-8045-f960c67a6936" Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.456807 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" event={"ID":"060e1c8d-dfa6-428f-bffe-d89ac3dab8c3","Type":"ContainerStarted","Data":"ff54c3cc814d92ee1cc321d45609ae8b43da463eafa4e73ee76a82ae916f552a"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.464595 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" event={"ID":"6245b73e-9fba-4ad7-bbbc-31db48c03825","Type":"ContainerStarted","Data":"7731d3601c99b3906aa086320f86fb3b157d3282fd194cb0c802ec8777cf95a8"} Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.466860 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" podUID="6245b73e-9fba-4ad7-bbbc-31db48c03825" Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.466915 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" event={"ID":"dd6346d8-9cf1-4364-b480-f4c2d872472f","Type":"ContainerStarted","Data":"0622fb49a57eae1e8e00740432fec781732d4906a697cdd1a79d78bf9f6d18c5"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.469519 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" event={"ID":"555ebb8f-1bc3-4b8d-9f37-cad92b48477c","Type":"ContainerStarted","Data":"70165325ea58298e820a24031f2133bcfafe721afd51512679019a903f8702c2"} Jan 24 00:18:19 crc kubenswrapper[4676]: E0124 00:18:19.470210 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:816d474f502d730d6a2522a272b0e09a2d579ac63617817655d60c54bda4191e\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" podUID="dd6346d8-9cf1-4364-b480-f4c2d872472f" Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.470647 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" event={"ID":"9a3f9a14-1138-425d-8a56-454b282d7d9f","Type":"ContainerStarted","Data":"ca37e0e8975c1fe950c6122163b3ae7d6c94476bcd0e2f65d36f111b4a09c40e"} Jan 24 00:18:19 crc kubenswrapper[4676]: I0124 00:18:19.473740 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" event={"ID":"53358678-d763-4b02-a157-86a57ebd0305","Type":"ContainerStarted","Data":"d2f11503b7cc79ec0c2b3bfdaa122f0a43cfe74be5d546eaf9d1087571849c60"} Jan 24 00:18:20 crc kubenswrapper[4676]: I0124 00:18:20.248103 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:20 crc kubenswrapper[4676]: E0124 00:18:20.248590 4676 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:20 crc kubenswrapper[4676]: E0124 00:18:20.248647 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert podName:921e121c-5261-4fe7-8171-6b634babedf4 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:24.248628938 +0000 UTC m=+888.278599949 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert") pod "infra-operator-controller-manager-58749ffdfb-jxx26" (UID: "921e121c-5261-4fe7-8171-6b634babedf4") : secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:20 crc kubenswrapper[4676]: E0124 00:18:20.492578 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:816d474f502d730d6a2522a272b0e09a2d579ac63617817655d60c54bda4191e\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" podUID="dd6346d8-9cf1-4364-b480-f4c2d872472f" Jan 24 00:18:20 crc kubenswrapper[4676]: E0124 00:18:20.493206 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:4e995cfa360a9d595a01b9c0541ab934692f2374203cb5738127dd784f793831\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" podUID="df9ab5f0-f577-4303-8045-f960c67a6936" Jan 24 00:18:20 crc kubenswrapper[4676]: E0124 00:18:20.493481 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" podUID="d85fa79d-818f-4079-aac4-f3fa51a90e9a" Jan 24 00:18:20 crc kubenswrapper[4676]: E0124 00:18:20.493939 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" podUID="6245b73e-9fba-4ad7-bbbc-31db48c03825" Jan 24 00:18:20 crc kubenswrapper[4676]: E0124 00:18:20.494622 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" podUID="fce4c8d0-b903-4873-8c89-2f4b9dd9c05d" Jan 24 00:18:20 crc kubenswrapper[4676]: I0124 00:18:20.862990 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:20 crc kubenswrapper[4676]: E0124 00:18:20.863167 4676 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:20 crc kubenswrapper[4676]: E0124 00:18:20.863238 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert podName:196f45b9-e656-4760-b058-e0b5c08a50d9 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:24.863221104 +0000 UTC m=+888.893192105 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" (UID: "196f45b9-e656-4760-b058-e0b5c08a50d9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:21 crc kubenswrapper[4676]: I0124 00:18:21.275950 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:21 crc kubenswrapper[4676]: I0124 00:18:21.276129 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:21 crc kubenswrapper[4676]: E0124 00:18:21.276135 4676 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 24 00:18:21 crc kubenswrapper[4676]: E0124 00:18:21.276231 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:25.276209336 +0000 UTC m=+889.306180337 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "webhook-server-cert" not found Jan 24 00:18:21 crc kubenswrapper[4676]: E0124 00:18:21.276284 4676 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 24 00:18:21 crc kubenswrapper[4676]: E0124 00:18:21.276364 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:25.27634521 +0000 UTC m=+889.306316211 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "metrics-server-cert" not found Jan 24 00:18:24 crc kubenswrapper[4676]: I0124 00:18:24.335901 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:24 crc kubenswrapper[4676]: E0124 00:18:24.336733 4676 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:24 crc kubenswrapper[4676]: E0124 00:18:24.336785 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert podName:921e121c-5261-4fe7-8171-6b634babedf4 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:32.336767287 +0000 UTC m=+896.366738288 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert") pod "infra-operator-controller-manager-58749ffdfb-jxx26" (UID: "921e121c-5261-4fe7-8171-6b634babedf4") : secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:24 crc kubenswrapper[4676]: I0124 00:18:24.944309 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:24 crc kubenswrapper[4676]: E0124 00:18:24.944577 4676 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:24 crc kubenswrapper[4676]: E0124 00:18:24.944669 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert podName:196f45b9-e656-4760-b058-e0b5c08a50d9 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:32.944645947 +0000 UTC m=+896.974616958 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" (UID: "196f45b9-e656-4760-b058-e0b5c08a50d9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:25 crc kubenswrapper[4676]: I0124 00:18:25.351030 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:25 crc kubenswrapper[4676]: I0124 00:18:25.351216 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:25 crc kubenswrapper[4676]: E0124 00:18:25.352009 4676 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 24 00:18:25 crc kubenswrapper[4676]: E0124 00:18:25.352092 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:33.352070258 +0000 UTC m=+897.382041269 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "webhook-server-cert" not found Jan 24 00:18:25 crc kubenswrapper[4676]: E0124 00:18:25.352492 4676 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 24 00:18:25 crc kubenswrapper[4676]: E0124 00:18:25.352584 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:33.352560403 +0000 UTC m=+897.382531434 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "metrics-server-cert" not found Jan 24 00:18:32 crc kubenswrapper[4676]: I0124 00:18:32.348538 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:32 crc kubenswrapper[4676]: E0124 00:18:32.348753 4676 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:32 crc kubenswrapper[4676]: E0124 00:18:32.349405 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert podName:921e121c-5261-4fe7-8171-6b634babedf4 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:48.349354671 +0000 UTC m=+912.379325712 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert") pod "infra-operator-controller-manager-58749ffdfb-jxx26" (UID: "921e121c-5261-4fe7-8171-6b634babedf4") : secret "infra-operator-webhook-server-cert" not found Jan 24 00:18:32 crc kubenswrapper[4676]: E0124 00:18:32.908811 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:f2035a0d3a8cc9434ab118078297f08cb8f3df98d1c75005279ee7915a3c2551" Jan 24 00:18:32 crc kubenswrapper[4676]: E0124 00:18:32.909629 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:f2035a0d3a8cc9434ab118078297f08cb8f3df98d1c75005279ee7915a3c2551,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmsrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6d9458688d-h6hzt_openstack-operators(ccb1ff12-bef7-4f23-b084-fae32f8202ac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:32 crc kubenswrapper[4676]: E0124 00:18:32.911001 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" podUID="ccb1ff12-bef7-4f23-b084-fae32f8202ac" Jan 24 00:18:32 crc kubenswrapper[4676]: I0124 00:18:32.959892 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:32 crc kubenswrapper[4676]: E0124 00:18:32.960349 4676 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:32 crc kubenswrapper[4676]: E0124 00:18:32.960492 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert podName:196f45b9-e656-4760-b058-e0b5c08a50d9 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:48.96046224 +0000 UTC m=+912.990433281 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" (UID: "196f45b9-e656-4760-b058-e0b5c08a50d9") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 24 00:18:33 crc kubenswrapper[4676]: I0124 00:18:33.366514 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:33 crc kubenswrapper[4676]: I0124 00:18:33.366668 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:33 crc kubenswrapper[4676]: E0124 00:18:33.366897 4676 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 24 00:18:33 crc kubenswrapper[4676]: E0124 00:18:33.366969 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:49.366946802 +0000 UTC m=+913.396917853 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "metrics-server-cert" not found Jan 24 00:18:33 crc kubenswrapper[4676]: E0124 00:18:33.367683 4676 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 24 00:18:33 crc kubenswrapper[4676]: E0124 00:18:33.367749 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs podName:9d79c791-c851-4c4a-aa2d-d175b668b0f5 nodeName:}" failed. No retries permitted until 2026-01-24 00:18:49.367730936 +0000 UTC m=+913.397701977 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs") pod "openstack-operator-controller-manager-5d5f8c4f48-md6zs" (UID: "9d79c791-c851-4c4a-aa2d-d175b668b0f5") : secret "webhook-server-cert" not found Jan 24 00:18:33 crc kubenswrapper[4676]: E0124 00:18:33.606418 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:f2035a0d3a8cc9434ab118078297f08cb8f3df98d1c75005279ee7915a3c2551\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" podUID="ccb1ff12-bef7-4f23-b084-fae32f8202ac" Jan 24 00:18:34 crc kubenswrapper[4676]: E0124 00:18:34.022930 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492" Jan 24 00:18:34 crc kubenswrapper[4676]: E0124 00:18:34.023237 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qsl2n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-594c8c9d5d-p2nr8_openstack-operators(5e9cf1cb-c413-45ad-8a51-bf35407fcdfe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:34 crc kubenswrapper[4676]: E0124 00:18:34.024776 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" podUID="5e9cf1cb-c413-45ad-8a51-bf35407fcdfe" Jan 24 00:18:34 crc kubenswrapper[4676]: E0124 00:18:34.611789 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492\\\"\"" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" podUID="5e9cf1cb-c413-45ad-8a51-bf35407fcdfe" Jan 24 00:18:36 crc kubenswrapper[4676]: E0124 00:18:36.491636 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:3311e627bcb860d9443592a2c67078417318c9eb77d8ef4d07f9aa7027d46822" Jan 24 00:18:36 crc kubenswrapper[4676]: E0124 00:18:36.492232 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:3311e627bcb860d9443592a2c67078417318c9eb77d8ef4d07f9aa7027d46822,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6f9hs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-77d5c5b54f-rbzsj_openstack-operators(6b8541f9-a37a-41d6-8006-3d0335c3abb5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:36 crc kubenswrapper[4676]: E0124 00:18:36.493452 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" podUID="6b8541f9-a37a-41d6-8006-3d0335c3abb5" Jan 24 00:18:36 crc kubenswrapper[4676]: E0124 00:18:36.630083 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:3311e627bcb860d9443592a2c67078417318c9eb77d8ef4d07f9aa7027d46822\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" podUID="6b8541f9-a37a-41d6-8006-3d0335c3abb5" Jan 24 00:18:40 crc kubenswrapper[4676]: E0124 00:18:40.993710 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:6c88312afa9673f7b72c558368034d7a488ead73080cdcdf581fe85b99263ece" Jan 24 00:18:40 crc kubenswrapper[4676]: E0124 00:18:40.994455 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:6c88312afa9673f7b72c558368034d7a488ead73080cdcdf581fe85b99263ece,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c2lp4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-b45d7bf98-js6db_openstack-operators(a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:40 crc kubenswrapper[4676]: E0124 00:18:40.995665 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" podUID="a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85" Jan 24 00:18:41 crc kubenswrapper[4676]: E0124 00:18:41.672167 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:6c88312afa9673f7b72c558368034d7a488ead73080cdcdf581fe85b99263ece\\\"\"" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" podUID="a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85" Jan 24 00:18:43 crc kubenswrapper[4676]: E0124 00:18:43.948025 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:9caae9b3ee328df678baa26454e45e47693acdadb27f9c635680597aaec43337" Jan 24 00:18:43 crc kubenswrapper[4676]: E0124 00:18:43.948725 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:9caae9b3ee328df678baa26454e45e47693acdadb27f9c635680597aaec43337,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-92kpk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-78fdd796fd-btxnv_openstack-operators(29e4b64d-19bd-419b-9e29-7a41e6f12ae0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:43 crc kubenswrapper[4676]: E0124 00:18:43.949958 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" podUID="29e4b64d-19bd-419b-9e29-7a41e6f12ae0" Jan 24 00:18:44 crc kubenswrapper[4676]: E0124 00:18:44.682047 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:9caae9b3ee328df678baa26454e45e47693acdadb27f9c635680597aaec43337\\\"\"" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" podUID="29e4b64d-19bd-419b-9e29-7a41e6f12ae0" Jan 24 00:18:46 crc kubenswrapper[4676]: E0124 00:18:46.744819 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:a8fc8f9d445b1232f446119015b226008b07c6a259f5bebc1fcbb39ec310afe5" Jan 24 00:18:46 crc kubenswrapper[4676]: E0124 00:18:46.745352 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:a8fc8f9d445b1232f446119015b226008b07c6a259f5bebc1fcbb39ec310afe5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-c5lfg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7bd9774b6-k8lw7_openstack-operators(ced74bcb-8345-40c5-b2d4-3d369f30b835): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:46 crc kubenswrapper[4676]: E0124 00:18:46.748828 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" podUID="ced74bcb-8345-40c5-b2d4-3d369f30b835" Jan 24 00:18:47 crc kubenswrapper[4676]: E0124 00:18:47.574132 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:4d55bd6418df3f63f4d3fe47bebf3f5498a520b3e14af98fe16c85ef9fd54d5e" Jan 24 00:18:47 crc kubenswrapper[4676]: E0124 00:18:47.574332 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:4d55bd6418df3f63f4d3fe47bebf3f5498a520b3e14af98fe16c85ef9fd54d5e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5fqcx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-598f7747c9-v25h4_openstack-operators(555ebb8f-1bc3-4b8d-9f37-cad92b48477c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:47 crc kubenswrapper[4676]: E0124 00:18:47.575540 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" podUID="555ebb8f-1bc3-4b8d-9f37-cad92b48477c" Jan 24 00:18:47 crc kubenswrapper[4676]: E0124 00:18:47.700278 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:4d55bd6418df3f63f4d3fe47bebf3f5498a520b3e14af98fe16c85ef9fd54d5e\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" podUID="555ebb8f-1bc3-4b8d-9f37-cad92b48477c" Jan 24 00:18:47 crc kubenswrapper[4676]: E0124 00:18:47.701570 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:a8fc8f9d445b1232f446119015b226008b07c6a259f5bebc1fcbb39ec310afe5\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" podUID="ced74bcb-8345-40c5-b2d4-3d369f30b835" Jan 24 00:18:48 crc kubenswrapper[4676]: E0124 00:18:48.244066 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8" Jan 24 00:18:48 crc kubenswrapper[4676]: E0124 00:18:48.244272 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l8zz7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-78c6999f6f-g4d8z_openstack-operators(dfc79179-d245-4360-be6e-8b43441e23ed): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:48 crc kubenswrapper[4676]: E0124 00:18:48.245494 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" podUID="dfc79179-d245-4360-be6e-8b43441e23ed" Jan 24 00:18:48 crc kubenswrapper[4676]: I0124 00:18:48.425144 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:48 crc kubenswrapper[4676]: I0124 00:18:48.430652 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/921e121c-5261-4fe7-8171-6b634babedf4-cert\") pod \"infra-operator-controller-manager-58749ffdfb-jxx26\" (UID: \"921e121c-5261-4fe7-8171-6b634babedf4\") " pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:48 crc kubenswrapper[4676]: I0124 00:18:48.547595 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-kcp2q" Jan 24 00:18:48 crc kubenswrapper[4676]: I0124 00:18:48.556661 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:18:48 crc kubenswrapper[4676]: E0124 00:18:48.705499 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" podUID="dfc79179-d245-4360-be6e-8b43441e23ed" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.034399 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.041091 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/196f45b9-e656-4760-b058-e0b5c08a50d9-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms\" (UID: \"196f45b9-e656-4760-b058-e0b5c08a50d9\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.340942 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rtqkf" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.350208 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.439901 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.440062 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.445432 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-metrics-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.446873 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/9d79c791-c851-4c4a-aa2d-d175b668b0f5-webhook-certs\") pod \"openstack-operator-controller-manager-5d5f8c4f48-md6zs\" (UID: \"9d79c791-c851-4c4a-aa2d-d175b668b0f5\") " pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.644593 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-2p4mv" Jan 24 00:18:49 crc kubenswrapper[4676]: I0124 00:18:49.650686 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:18:51 crc kubenswrapper[4676]: E0124 00:18:51.270428 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf" Jan 24 00:18:51 crc kubenswrapper[4676]: E0124 00:18:51.272095 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wqqxj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-55db956ddc-49nrq_openstack-operators(53358678-d763-4b02-a157-86a57ebd0305): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:51 crc kubenswrapper[4676]: E0124 00:18:51.273416 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" podUID="53358678-d763-4b02-a157-86a57ebd0305" Jan 24 00:18:51 crc kubenswrapper[4676]: E0124 00:18:51.726544 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:8b3bfb9e86618b7ac69443939b0968fae28a22cd62ea1e429b599ff9f8a5f8cf\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" podUID="53358678-d763-4b02-a157-86a57ebd0305" Jan 24 00:18:52 crc kubenswrapper[4676]: E0124 00:18:52.013932 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d" Jan 24 00:18:52 crc kubenswrapper[4676]: E0124 00:18:52.014122 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2xd87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-69797bbcbd-gqg82_openstack-operators(6245b73e-9fba-4ad7-bbbc-31db48c03825): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:52 crc kubenswrapper[4676]: E0124 00:18:52.015520 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" podUID="6245b73e-9fba-4ad7-bbbc-31db48c03825" Jan 24 00:18:52 crc kubenswrapper[4676]: E0124 00:18:52.595107 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349" Jan 24 00:18:52 crc kubenswrapper[4676]: E0124 00:18:52.595299 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-np4zv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b8b6d4659-2gzqg_openstack-operators(9a3f9a14-1138-425d-8a56-454b282d7d9f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:52 crc kubenswrapper[4676]: E0124 00:18:52.596824 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" podUID="9a3f9a14-1138-425d-8a56-454b282d7d9f" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.632327 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8rvgb"] Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.635127 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.643637 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8rvgb"] Jan 24 00:18:52 crc kubenswrapper[4676]: E0124 00:18:52.737191 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" podUID="9a3f9a14-1138-425d-8a56-454b282d7d9f" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.792998 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-catalog-content\") pod \"community-operators-8rvgb\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.793145 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-utilities\") pod \"community-operators-8rvgb\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.793240 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sf2s\" (UniqueName: \"kubernetes.io/projected/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-kube-api-access-4sf2s\") pod \"community-operators-8rvgb\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.894118 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sf2s\" (UniqueName: \"kubernetes.io/projected/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-kube-api-access-4sf2s\") pod \"community-operators-8rvgb\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.894183 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-catalog-content\") pod \"community-operators-8rvgb\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.894223 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-utilities\") pod \"community-operators-8rvgb\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.894699 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-utilities\") pod \"community-operators-8rvgb\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.894808 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-catalog-content\") pod \"community-operators-8rvgb\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.920033 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sf2s\" (UniqueName: \"kubernetes.io/projected/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-kube-api-access-4sf2s\") pod \"community-operators-8rvgb\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:52 crc kubenswrapper[4676]: I0124 00:18:52.963266 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:18:54 crc kubenswrapper[4676]: E0124 00:18:54.634286 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.69:5001/openstack-k8s-operators/swift-operator:23f4c2121362f560ce87ad957773c006d47a5413" Jan 24 00:18:54 crc kubenswrapper[4676]: E0124 00:18:54.634865 4676 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.69:5001/openstack-k8s-operators/swift-operator:23f4c2121362f560ce87ad957773c006d47a5413" Jan 24 00:18:54 crc kubenswrapper[4676]: E0124 00:18:54.635026 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.69:5001/openstack-k8s-operators/swift-operator:23f4c2121362f560ce87ad957773c006d47a5413,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8j5xn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5df95d5965-h8wx9_openstack-operators(b0c8972b-31d7-40c1-bc65-1478718d41a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:54 crc kubenswrapper[4676]: E0124 00:18:54.636152 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" podUID="b0c8972b-31d7-40c1-bc65-1478718d41a5" Jan 24 00:18:54 crc kubenswrapper[4676]: E0124 00:18:54.746417 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.69:5001/openstack-k8s-operators/swift-operator:23f4c2121362f560ce87ad957773c006d47a5413\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" podUID="b0c8972b-31d7-40c1-bc65-1478718d41a5" Jan 24 00:18:55 crc kubenswrapper[4676]: E0124 00:18:55.588873 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:816d474f502d730d6a2522a272b0e09a2d579ac63617817655d60c54bda4191e" Jan 24 00:18:55 crc kubenswrapper[4676]: E0124 00:18:55.589173 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:816d474f502d730d6a2522a272b0e09a2d579ac63617817655d60c54bda4191e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-22p7t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-78d58447c5-wqbcz_openstack-operators(dd6346d8-9cf1-4364-b480-f4c2d872472f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:55 crc kubenswrapper[4676]: E0124 00:18:55.591074 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" podUID="dd6346d8-9cf1-4364-b480-f4c2d872472f" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.194279 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lbhrz"] Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.196217 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.205632 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbhrz"] Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.376686 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksl7d\" (UniqueName: \"kubernetes.io/projected/6fb6d395-75c7-4f58-85cf-a97ad12bb288-kube-api-access-ksl7d\") pod \"redhat-marketplace-lbhrz\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.376809 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-utilities\") pod \"redhat-marketplace-lbhrz\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.376834 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-catalog-content\") pod \"redhat-marketplace-lbhrz\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.478362 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksl7d\" (UniqueName: \"kubernetes.io/projected/6fb6d395-75c7-4f58-85cf-a97ad12bb288-kube-api-access-ksl7d\") pod \"redhat-marketplace-lbhrz\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.478538 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-utilities\") pod \"redhat-marketplace-lbhrz\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.478572 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-catalog-content\") pod \"redhat-marketplace-lbhrz\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.479315 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-catalog-content\") pod \"redhat-marketplace-lbhrz\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.479861 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-utilities\") pod \"redhat-marketplace-lbhrz\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.502349 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksl7d\" (UniqueName: \"kubernetes.io/projected/6fb6d395-75c7-4f58-85cf-a97ad12bb288-kube-api-access-ksl7d\") pod \"redhat-marketplace-lbhrz\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:57 crc kubenswrapper[4676]: I0124 00:18:57.521479 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:18:59 crc kubenswrapper[4676]: E0124 00:18:59.929069 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0" Jan 24 00:18:59 crc kubenswrapper[4676]: E0124 00:18:59.929530 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4pggf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5d646b7d76-tdqwh_openstack-operators(d85fa79d-818f-4079-aac4-f3fa51a90e9a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:18:59 crc kubenswrapper[4676]: E0124 00:18:59.930747 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" podUID="d85fa79d-818f-4079-aac4-f3fa51a90e9a" Jan 24 00:19:00 crc kubenswrapper[4676]: E0124 00:19:00.634208 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:4e995cfa360a9d595a01b9c0541ab934692f2374203cb5738127dd784f793831" Jan 24 00:19:00 crc kubenswrapper[4676]: E0124 00:19:00.634390 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:4e995cfa360a9d595a01b9c0541ab934692f2374203cb5738127dd784f793831,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l8pmb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-6b8bc8d87d-4xp45_openstack-operators(df9ab5f0-f577-4303-8045-f960c67a6936): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:19:00 crc kubenswrapper[4676]: E0124 00:19:00.635779 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" podUID="df9ab5f0-f577-4303-8045-f960c67a6936" Jan 24 00:19:01 crc kubenswrapper[4676]: E0124 00:19:01.295891 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Jan 24 00:19:01 crc kubenswrapper[4676]: E0124 00:19:01.296317 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8ffff,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-h44qw_openstack-operators(fce4c8d0-b903-4873-8c89-2f4b9dd9c05d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:19:01 crc kubenswrapper[4676]: E0124 00:19:01.297592 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" podUID="fce4c8d0-b903-4873-8c89-2f4b9dd9c05d" Jan 24 00:19:01 crc kubenswrapper[4676]: I0124 00:19:01.785225 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs"] Jan 24 00:19:01 crc kubenswrapper[4676]: I0124 00:19:01.920134 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26"] Jan 24 00:19:01 crc kubenswrapper[4676]: I0124 00:19:01.938663 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms"] Jan 24 00:19:01 crc kubenswrapper[4676]: W0124 00:19:01.986563 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod196f45b9_e656_4760_b058_e0b5c08a50d9.slice/crio-c719a5059cec3d49dd2a99d136341fb5121ae1ac6a487e244576edb0fad3b540 WatchSource:0}: Error finding container c719a5059cec3d49dd2a99d136341fb5121ae1ac6a487e244576edb0fad3b540: Status 404 returned error can't find the container with id c719a5059cec3d49dd2a99d136341fb5121ae1ac6a487e244576edb0fad3b540 Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.101917 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8rvgb"] Jan 24 00:19:02 crc kubenswrapper[4676]: W0124 00:19:02.119479 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddccfed4b_e148_4e09_8c96_f5a28fefd6b7.slice/crio-ae7b958b56efd4b4607c5292b0d29fa3ac42c0d3302dd61deaf7d117c5fd353a WatchSource:0}: Error finding container ae7b958b56efd4b4607c5292b0d29fa3ac42c0d3302dd61deaf7d117c5fd353a: Status 404 returned error can't find the container with id ae7b958b56efd4b4607c5292b0d29fa3ac42c0d3302dd61deaf7d117c5fd353a Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.132446 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbhrz"] Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.827448 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" event={"ID":"29e4b64d-19bd-419b-9e29-7a41e6f12ae0","Type":"ContainerStarted","Data":"c94ff148de8e830b54fedc7e8ea23b763f7e15f45a9e67d7d1b8f51192ef4cc5"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.828517 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.829630 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" event={"ID":"5e9cf1cb-c413-45ad-8a51-bf35407fcdfe","Type":"ContainerStarted","Data":"79aed58d1500a29487f2ad99122f300fe0d4be30fe9544de12a9b1664479653a"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.829959 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.831241 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rvgb" event={"ID":"dccfed4b-e148-4e09-8c96-f5a28fefd6b7","Type":"ContainerStarted","Data":"ae7b958b56efd4b4607c5292b0d29fa3ac42c0d3302dd61deaf7d117c5fd353a"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.832423 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbhrz" event={"ID":"6fb6d395-75c7-4f58-85cf-a97ad12bb288","Type":"ContainerStarted","Data":"dc8c29c9b6a4f98617ba640622dff3ce5e2bbf9034752fb810b692b938fe4d40"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.833865 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" event={"ID":"a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85","Type":"ContainerStarted","Data":"edd1392f87f77e9c36877306b88f09986ff4d5a5728c278d969a1d3b2dae468a"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.834256 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.835460 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" event={"ID":"9d79c791-c851-4c4a-aa2d-d175b668b0f5","Type":"ContainerStarted","Data":"e03878e2d87db1a4615d65b136e9d1a03b12224c279491cec4309bd6c2754e99"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.835483 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" event={"ID":"9d79c791-c851-4c4a-aa2d-d175b668b0f5","Type":"ContainerStarted","Data":"36b786131833cb15f0894c7ec6c55fca661bb1b7d3c23584010803300759f98a"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.835805 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.836826 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" event={"ID":"6b8541f9-a37a-41d6-8006-3d0335c3abb5","Type":"ContainerStarted","Data":"c27d94b2d74a179ee31054a5427091d0cc5a482a5eba9d918140283913ca1c9c"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.837140 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.838518 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" event={"ID":"ccb1ff12-bef7-4f23-b084-fae32f8202ac","Type":"ContainerStarted","Data":"225be960d85ba3f02a2da8cb559d60253fe2af3074ac37328062999f50d2800c"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.838831 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.840060 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" event={"ID":"555ebb8f-1bc3-4b8d-9f37-cad92b48477c","Type":"ContainerStarted","Data":"5d6d33e8954d6997cc050ec656a057d92cf8476f658755a78fb8d06a91e7ad20"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.840289 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.841576 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" event={"ID":"02123851-7d2f-477b-9c60-5a9922a0bc97","Type":"ContainerStarted","Data":"cc5aac30db056a91920b0f85abbe608ee3a61b6063fd40d80b2dc0f5cffa8ba4"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.841698 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.842320 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" event={"ID":"196f45b9-e656-4760-b058-e0b5c08a50d9","Type":"ContainerStarted","Data":"c719a5059cec3d49dd2a99d136341fb5121ae1ac6a487e244576edb0fad3b540"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.843683 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" event={"ID":"4ce661f6-26e2-4da2-a759-e493a60587b2","Type":"ContainerStarted","Data":"4f123b889d5c8f17006a5118c2e0a95be43d36b1800fd499c154a705f93959c7"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.844096 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.845121 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" event={"ID":"921e121c-5261-4fe7-8171-6b634babedf4","Type":"ContainerStarted","Data":"4cd646ceb9964735b20736b129c5849347664e8cad6dc534ad2a31cae9a4e1c6"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.846209 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" event={"ID":"0cd05b9f-6699-46e3-ae36-9f21352e6c8e","Type":"ContainerStarted","Data":"247343947e75491a04b4468dc003f26c722cb2cf09a2a6f4c36d0077d4814257"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.846557 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.847585 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" event={"ID":"ced74bcb-8345-40c5-b2d4-3d369f30b835","Type":"ContainerStarted","Data":"7857adbd42df4c3bb3e0e71e201ae55278800fb5d39ae8bac371145af179f2b3"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.847920 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.850064 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" event={"ID":"060e1c8d-dfa6-428f-bffe-d89ac3dab8c3","Type":"ContainerStarted","Data":"8c070fc6f5bbc004f03ef8f9daf5e548d66db0e1aaf71692113d4beb7502ea1e"} Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.850368 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.923305 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" podStartSLOduration=45.923289679 podStartE2EDuration="45.923289679s" podCreationTimestamp="2026-01-24 00:18:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:19:02.918956236 +0000 UTC m=+926.948927237" watchObservedRunningTime="2026-01-24 00:19:02.923289679 +0000 UTC m=+926.953260680" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.924483 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" podStartSLOduration=3.822036786 podStartE2EDuration="46.924475656s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.291600356 +0000 UTC m=+882.321571367" lastFinishedPulling="2026-01-24 00:19:01.394039236 +0000 UTC m=+925.424010237" observedRunningTime="2026-01-24 00:19:02.8636002 +0000 UTC m=+926.893571201" watchObservedRunningTime="2026-01-24 00:19:02.924475656 +0000 UTC m=+926.954446657" Jan 24 00:19:02 crc kubenswrapper[4676]: I0124 00:19:02.952207 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" podStartSLOduration=4.313833147 podStartE2EDuration="46.952187331s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.650724236 +0000 UTC m=+882.680695237" lastFinishedPulling="2026-01-24 00:19:01.28907842 +0000 UTC m=+925.319049421" observedRunningTime="2026-01-24 00:19:02.947918559 +0000 UTC m=+926.977889560" watchObservedRunningTime="2026-01-24 00:19:02.952187331 +0000 UTC m=+926.982158332" Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.117870 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" podStartSLOduration=4.578457924 podStartE2EDuration="46.117852998s" podCreationTimestamp="2026-01-24 00:18:17 +0000 UTC" firstStartedPulling="2026-01-24 00:18:19.081050953 +0000 UTC m=+883.111021954" lastFinishedPulling="2026-01-24 00:19:00.620446017 +0000 UTC m=+924.650417028" observedRunningTime="2026-01-24 00:19:03.056827187 +0000 UTC m=+927.086798188" watchObservedRunningTime="2026-01-24 00:19:03.117852998 +0000 UTC m=+927.147823999" Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.176690 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" podStartSLOduration=4.407043971 podStartE2EDuration="47.176671961s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.626756128 +0000 UTC m=+882.656727129" lastFinishedPulling="2026-01-24 00:19:01.396384118 +0000 UTC m=+925.426355119" observedRunningTime="2026-01-24 00:19:03.119315993 +0000 UTC m=+927.149286994" watchObservedRunningTime="2026-01-24 00:19:03.176671961 +0000 UTC m=+927.206642962" Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.209624 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" podStartSLOduration=4.811066826 podStartE2EDuration="47.209606717s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.223138455 +0000 UTC m=+882.253109456" lastFinishedPulling="2026-01-24 00:19:00.621678336 +0000 UTC m=+924.651649347" observedRunningTime="2026-01-24 00:19:03.180880671 +0000 UTC m=+927.210851672" watchObservedRunningTime="2026-01-24 00:19:03.209606717 +0000 UTC m=+927.239577718" Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.212228 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" podStartSLOduration=4.286507776 podStartE2EDuration="47.212222587s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.653968997 +0000 UTC m=+882.683939998" lastFinishedPulling="2026-01-24 00:19:01.579683808 +0000 UTC m=+925.609654809" observedRunningTime="2026-01-24 00:19:03.208494383 +0000 UTC m=+927.238465384" watchObservedRunningTime="2026-01-24 00:19:03.212222587 +0000 UTC m=+927.242193588" Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.243644 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" podStartSLOduration=4.291225612 podStartE2EDuration="47.243630456s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.616685878 +0000 UTC m=+882.646656879" lastFinishedPulling="2026-01-24 00:19:01.569090722 +0000 UTC m=+925.599061723" observedRunningTime="2026-01-24 00:19:03.240353684 +0000 UTC m=+927.270324685" watchObservedRunningTime="2026-01-24 00:19:03.243630456 +0000 UTC m=+927.273601457" Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.258598 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.304437 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" podStartSLOduration=13.269747103 podStartE2EDuration="47.304423619s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.546906077 +0000 UTC m=+882.576877078" lastFinishedPulling="2026-01-24 00:18:52.581582593 +0000 UTC m=+916.611553594" observedRunningTime="2026-01-24 00:19:03.301219151 +0000 UTC m=+927.331190152" watchObservedRunningTime="2026-01-24 00:19:03.304423619 +0000 UTC m=+927.334394620" Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.343146 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" podStartSLOduration=12.530652975 podStartE2EDuration="46.343129232s" podCreationTimestamp="2026-01-24 00:18:17 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.77049988 +0000 UTC m=+882.800470871" lastFinishedPulling="2026-01-24 00:18:52.582976127 +0000 UTC m=+916.612947128" observedRunningTime="2026-01-24 00:19:03.335503628 +0000 UTC m=+927.365474629" watchObservedRunningTime="2026-01-24 00:19:03.343129232 +0000 UTC m=+927.373100233" Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.377230 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" podStartSLOduration=12.728408723 podStartE2EDuration="47.377215533s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:17.933724453 +0000 UTC m=+881.963695444" lastFinishedPulling="2026-01-24 00:18:52.582531253 +0000 UTC m=+916.612502254" observedRunningTime="2026-01-24 00:19:03.373954773 +0000 UTC m=+927.403925774" watchObservedRunningTime="2026-01-24 00:19:03.377215533 +0000 UTC m=+927.407186534" Jan 24 00:19:03 crc kubenswrapper[4676]: I0124 00:19:03.414873 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" podStartSLOduration=17.280900119000002 podStartE2EDuration="47.414857184s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.092035003 +0000 UTC m=+882.122006004" lastFinishedPulling="2026-01-24 00:18:48.225992068 +0000 UTC m=+912.255963069" observedRunningTime="2026-01-24 00:19:03.411760099 +0000 UTC m=+927.441731100" watchObservedRunningTime="2026-01-24 00:19:03.414857184 +0000 UTC m=+927.444828185" Jan 24 00:19:04 crc kubenswrapper[4676]: I0124 00:19:04.870175 4676 generic.go:334] "Generic (PLEG): container finished" podID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerID="8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943" exitCode=0 Jan 24 00:19:04 crc kubenswrapper[4676]: I0124 00:19:04.870339 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rvgb" event={"ID":"dccfed4b-e148-4e09-8c96-f5a28fefd6b7","Type":"ContainerDied","Data":"8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943"} Jan 24 00:19:05 crc kubenswrapper[4676]: E0124 00:19:05.267310 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:c8dde42dafd41026ed2e4cfc26efc0fff63c4ba9d31326ae7dc644ccceaafa9d\\\"\"" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" podUID="6245b73e-9fba-4ad7-bbbc-31db48c03825" Jan 24 00:19:05 crc kubenswrapper[4676]: I0124 00:19:05.884081 4676 generic.go:334] "Generic (PLEG): container finished" podID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerID="fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49" exitCode=0 Jan 24 00:19:05 crc kubenswrapper[4676]: I0124 00:19:05.884452 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbhrz" event={"ID":"6fb6d395-75c7-4f58-85cf-a97ad12bb288","Type":"ContainerDied","Data":"fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49"} Jan 24 00:19:05 crc kubenswrapper[4676]: I0124 00:19:05.886971 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" event={"ID":"9a3f9a14-1138-425d-8a56-454b282d7d9f","Type":"ContainerStarted","Data":"78680acd5bedfe77b85e6fb17f9b64dce6f8a3e4764ff9f094cfecf560d8e821"} Jan 24 00:19:05 crc kubenswrapper[4676]: I0124 00:19:05.887615 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" Jan 24 00:19:05 crc kubenswrapper[4676]: I0124 00:19:05.902851 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" event={"ID":"dfc79179-d245-4360-be6e-8b43441e23ed","Type":"ContainerStarted","Data":"3e3dcb284379f3f1783412650f16694ff890f9172e38e5b94ec318f1d6c89e9c"} Jan 24 00:19:05 crc kubenswrapper[4676]: I0124 00:19:05.903312 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" Jan 24 00:19:05 crc kubenswrapper[4676]: I0124 00:19:05.917289 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" podStartSLOduration=3.2913077250000002 podStartE2EDuration="49.917268499s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.647193508 +0000 UTC m=+882.677164509" lastFinishedPulling="2026-01-24 00:19:05.273154282 +0000 UTC m=+929.303125283" observedRunningTime="2026-01-24 00:19:05.912815142 +0000 UTC m=+929.942786143" watchObservedRunningTime="2026-01-24 00:19:05.917268499 +0000 UTC m=+929.947239490" Jan 24 00:19:05 crc kubenswrapper[4676]: I0124 00:19:05.941476 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" podStartSLOduration=3.350687125 podStartE2EDuration="49.941461115s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.674632504 +0000 UTC m=+882.704603505" lastFinishedPulling="2026-01-24 00:19:05.265406484 +0000 UTC m=+929.295377495" observedRunningTime="2026-01-24 00:19:05.93706612 +0000 UTC m=+929.967037111" watchObservedRunningTime="2026-01-24 00:19:05.941461115 +0000 UTC m=+929.971432116" Jan 24 00:19:06 crc kubenswrapper[4676]: I0124 00:19:06.592427 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-69cf5d4557-c8c6m" Jan 24 00:19:06 crc kubenswrapper[4676]: I0124 00:19:06.670167 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-js6db" Jan 24 00:19:06 crc kubenswrapper[4676]: I0124 00:19:06.704874 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-p2nr8" Jan 24 00:19:06 crc kubenswrapper[4676]: I0124 00:19:06.777521 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-rbzsj" Jan 24 00:19:06 crc kubenswrapper[4676]: I0124 00:19:06.915490 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" event={"ID":"53358678-d763-4b02-a157-86a57ebd0305","Type":"ContainerStarted","Data":"949f9736c183d87736871cec1ab379c52ad00689b92be4b99d7a8a69463cb60a"} Jan 24 00:19:06 crc kubenswrapper[4676]: I0124 00:19:06.915876 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" Jan 24 00:19:06 crc kubenswrapper[4676]: I0124 00:19:06.932952 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" podStartSLOduration=3.8781462060000003 podStartE2EDuration="50.93293627s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.647362124 +0000 UTC m=+882.677333125" lastFinishedPulling="2026-01-24 00:19:05.702152188 +0000 UTC m=+929.732123189" observedRunningTime="2026-01-24 00:19:06.931264349 +0000 UTC m=+930.961235350" watchObservedRunningTime="2026-01-24 00:19:06.93293627 +0000 UTC m=+930.962907271" Jan 24 00:19:07 crc kubenswrapper[4676]: I0124 00:19:07.055061 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-598f7747c9-v25h4" Jan 24 00:19:07 crc kubenswrapper[4676]: I0124 00:19:07.085927 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n" Jan 24 00:19:07 crc kubenswrapper[4676]: I0124 00:19:07.444660 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7bd9774b6-k8lw7" Jan 24 00:19:07 crc kubenswrapper[4676]: I0124 00:19:07.704043 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-4c5zl" Jan 24 00:19:07 crc kubenswrapper[4676]: I0124 00:19:07.767027 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6d9458688d-h6hzt" Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.927402 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" event={"ID":"b0c8972b-31d7-40c1-bc65-1478718d41a5","Type":"ContainerStarted","Data":"abfef511ba5d3692f30535b0933f2886ff27a0633dd8cc09c8c3f685ca22fbc8"} Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.927976 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.931355 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" event={"ID":"921e121c-5261-4fe7-8171-6b634babedf4","Type":"ContainerStarted","Data":"a93d80bd6a6ba10512cd156d0877df7f548f4aabc39afaf8b53c2a37fa95c23d"} Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.931645 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.933991 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rvgb" event={"ID":"dccfed4b-e148-4e09-8c96-f5a28fefd6b7","Type":"ContainerStarted","Data":"d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff"} Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.936034 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbhrz" event={"ID":"6fb6d395-75c7-4f58-85cf-a97ad12bb288","Type":"ContainerStarted","Data":"09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04"} Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.937512 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" event={"ID":"196f45b9-e656-4760-b058-e0b5c08a50d9","Type":"ContainerStarted","Data":"37cb2962b3de94522bfdf9940464761e018897a697a1edfbd7e0ae5fdfb20c22"} Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.937891 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.947087 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" podStartSLOduration=3.167566952 podStartE2EDuration="52.947070033s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.7792613 +0000 UTC m=+882.809232301" lastFinishedPulling="2026-01-24 00:19:08.558764341 +0000 UTC m=+932.588735382" observedRunningTime="2026-01-24 00:19:08.943236854 +0000 UTC m=+932.973207865" watchObservedRunningTime="2026-01-24 00:19:08.947070033 +0000 UTC m=+932.977041034" Jan 24 00:19:08 crc kubenswrapper[4676]: I0124 00:19:08.990482 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" podStartSLOduration=46.486801013 podStartE2EDuration="52.99046723s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:19:01.99492693 +0000 UTC m=+926.024897931" lastFinishedPulling="2026-01-24 00:19:08.498593137 +0000 UTC m=+932.528564148" observedRunningTime="2026-01-24 00:19:08.983130314 +0000 UTC m=+933.013101315" watchObservedRunningTime="2026-01-24 00:19:08.99046723 +0000 UTC m=+933.020438221" Jan 24 00:19:09 crc kubenswrapper[4676]: I0124 00:19:09.008811 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" podStartSLOduration=46.480445398 podStartE2EDuration="53.008792066s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:19:01.964409519 +0000 UTC m=+925.994380520" lastFinishedPulling="2026-01-24 00:19:08.492756177 +0000 UTC m=+932.522727188" observedRunningTime="2026-01-24 00:19:08.999278632 +0000 UTC m=+933.029249623" watchObservedRunningTime="2026-01-24 00:19:09.008792066 +0000 UTC m=+933.038763067" Jan 24 00:19:09 crc kubenswrapper[4676]: I0124 00:19:09.363833 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:19:09 crc kubenswrapper[4676]: I0124 00:19:09.363887 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:19:09 crc kubenswrapper[4676]: I0124 00:19:09.655741 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5d5f8c4f48-md6zs" Jan 24 00:19:10 crc kubenswrapper[4676]: E0124 00:19:10.256690 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:816d474f502d730d6a2522a272b0e09a2d579ac63617817655d60c54bda4191e\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" podUID="dd6346d8-9cf1-4364-b480-f4c2d872472f" Jan 24 00:19:11 crc kubenswrapper[4676]: I0124 00:19:11.959011 4676 generic.go:334] "Generic (PLEG): container finished" podID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerID="d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff" exitCode=0 Jan 24 00:19:11 crc kubenswrapper[4676]: I0124 00:19:11.959109 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rvgb" event={"ID":"dccfed4b-e148-4e09-8c96-f5a28fefd6b7","Type":"ContainerDied","Data":"d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff"} Jan 24 00:19:11 crc kubenswrapper[4676]: I0124 00:19:11.963431 4676 generic.go:334] "Generic (PLEG): container finished" podID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerID="09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04" exitCode=0 Jan 24 00:19:11 crc kubenswrapper[4676]: I0124 00:19:11.963503 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbhrz" event={"ID":"6fb6d395-75c7-4f58-85cf-a97ad12bb288","Type":"ContainerDied","Data":"09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04"} Jan 24 00:19:12 crc kubenswrapper[4676]: E0124 00:19:12.257054 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:4e995cfa360a9d595a01b9c0541ab934692f2374203cb5738127dd784f793831\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" podUID="df9ab5f0-f577-4303-8045-f960c67a6936" Jan 24 00:19:12 crc kubenswrapper[4676]: E0124 00:19:12.257474 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" podUID="fce4c8d0-b903-4873-8c89-2f4b9dd9c05d" Jan 24 00:19:13 crc kubenswrapper[4676]: I0124 00:19:13.983245 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rvgb" event={"ID":"dccfed4b-e148-4e09-8c96-f5a28fefd6b7","Type":"ContainerStarted","Data":"055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1"} Jan 24 00:19:13 crc kubenswrapper[4676]: I0124 00:19:13.985436 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbhrz" event={"ID":"6fb6d395-75c7-4f58-85cf-a97ad12bb288","Type":"ContainerStarted","Data":"a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea"} Jan 24 00:19:14 crc kubenswrapper[4676]: I0124 00:19:14.034500 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8rvgb" podStartSLOduration=15.113324433 podStartE2EDuration="22.034487149s" podCreationTimestamp="2026-01-24 00:18:52 +0000 UTC" firstStartedPulling="2026-01-24 00:19:05.904332531 +0000 UTC m=+929.934303532" lastFinishedPulling="2026-01-24 00:19:12.825495247 +0000 UTC m=+936.855466248" observedRunningTime="2026-01-24 00:19:14.033719025 +0000 UTC m=+938.063690026" watchObservedRunningTime="2026-01-24 00:19:14.034487149 +0000 UTC m=+938.064458150" Jan 24 00:19:14 crc kubenswrapper[4676]: I0124 00:19:14.076975 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lbhrz" podStartSLOduration=10.150678542 podStartE2EDuration="17.076957487s" podCreationTimestamp="2026-01-24 00:18:57 +0000 UTC" firstStartedPulling="2026-01-24 00:19:05.885496299 +0000 UTC m=+929.915467300" lastFinishedPulling="2026-01-24 00:19:12.811775244 +0000 UTC m=+936.841746245" observedRunningTime="2026-01-24 00:19:14.071934943 +0000 UTC m=+938.101905954" watchObservedRunningTime="2026-01-24 00:19:14.076957487 +0000 UTC m=+938.106928488" Jan 24 00:19:15 crc kubenswrapper[4676]: E0124 00:19:15.257623 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:65cfe5b9d5b0571aaf8ff9840b12cc56e90ca4cef162dd260c3a9fa2b52c6dd0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" podUID="d85fa79d-818f-4079-aac4-f3fa51a90e9a" Jan 24 00:19:16 crc kubenswrapper[4676]: I0124 00:19:16.569553 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-fgplq" Jan 24 00:19:16 crc kubenswrapper[4676]: I0124 00:19:16.664648 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-btxnv" Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.006323 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" event={"ID":"6245b73e-9fba-4ad7-bbbc-31db48c03825","Type":"ContainerStarted","Data":"25b8ffc3686fc2ad6e94b92751d7a382f2ddf94fa692fbc5f2bdc927b0241a35"} Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.007490 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.026105 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" podStartSLOduration=2.416502544 podStartE2EDuration="1m0.026081944s" podCreationTimestamp="2026-01-24 00:18:17 +0000 UTC" firstStartedPulling="2026-01-24 00:18:19.102109652 +0000 UTC m=+883.132080653" lastFinishedPulling="2026-01-24 00:19:16.711689052 +0000 UTC m=+940.741660053" observedRunningTime="2026-01-24 00:19:17.022614617 +0000 UTC m=+941.052585628" watchObservedRunningTime="2026-01-24 00:19:17.026081944 +0000 UTC m=+941.056052955" Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.030623 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-2gzqg" Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.042959 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-g4d8z" Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.296584 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-55db956ddc-49nrq" Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.522401 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.522445 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.587509 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:19:17 crc kubenswrapper[4676]: I0124 00:19:17.633433 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5df95d5965-h8wx9" Jan 24 00:19:18 crc kubenswrapper[4676]: I0124 00:19:18.089564 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:19:18 crc kubenswrapper[4676]: I0124 00:19:18.136426 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbhrz"] Jan 24 00:19:18 crc kubenswrapper[4676]: I0124 00:19:18.563724 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-58749ffdfb-jxx26" Jan 24 00:19:19 crc kubenswrapper[4676]: I0124 00:19:19.358145 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms" Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.028252 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lbhrz" podUID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerName="registry-server" containerID="cri-o://a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea" gracePeriod=2 Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.456251 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.612612 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-utilities\") pod \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.612693 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksl7d\" (UniqueName: \"kubernetes.io/projected/6fb6d395-75c7-4f58-85cf-a97ad12bb288-kube-api-access-ksl7d\") pod \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.612763 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-catalog-content\") pod \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\" (UID: \"6fb6d395-75c7-4f58-85cf-a97ad12bb288\") " Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.613557 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-utilities" (OuterVolumeSpecName: "utilities") pod "6fb6d395-75c7-4f58-85cf-a97ad12bb288" (UID: "6fb6d395-75c7-4f58-85cf-a97ad12bb288"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.621675 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fb6d395-75c7-4f58-85cf-a97ad12bb288-kube-api-access-ksl7d" (OuterVolumeSpecName: "kube-api-access-ksl7d") pod "6fb6d395-75c7-4f58-85cf-a97ad12bb288" (UID: "6fb6d395-75c7-4f58-85cf-a97ad12bb288"). InnerVolumeSpecName "kube-api-access-ksl7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.638274 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6fb6d395-75c7-4f58-85cf-a97ad12bb288" (UID: "6fb6d395-75c7-4f58-85cf-a97ad12bb288"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.714940 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.714973 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksl7d\" (UniqueName: \"kubernetes.io/projected/6fb6d395-75c7-4f58-85cf-a97ad12bb288-kube-api-access-ksl7d\") on node \"crc\" DevicePath \"\"" Jan 24 00:19:20 crc kubenswrapper[4676]: I0124 00:19:20.714984 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6fb6d395-75c7-4f58-85cf-a97ad12bb288-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.036827 4676 generic.go:334] "Generic (PLEG): container finished" podID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerID="a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea" exitCode=0 Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.036871 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbhrz" event={"ID":"6fb6d395-75c7-4f58-85cf-a97ad12bb288","Type":"ContainerDied","Data":"a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea"} Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.036909 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbhrz" event={"ID":"6fb6d395-75c7-4f58-85cf-a97ad12bb288","Type":"ContainerDied","Data":"dc8c29c9b6a4f98617ba640622dff3ce5e2bbf9034752fb810b692b938fe4d40"} Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.036915 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbhrz" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.036929 4676 scope.go:117] "RemoveContainer" containerID="a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.069759 4676 scope.go:117] "RemoveContainer" containerID="09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.076087 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbhrz"] Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.079387 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbhrz"] Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.086293 4676 scope.go:117] "RemoveContainer" containerID="fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.105515 4676 scope.go:117] "RemoveContainer" containerID="a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea" Jan 24 00:19:21 crc kubenswrapper[4676]: E0124 00:19:21.105860 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea\": container with ID starting with a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea not found: ID does not exist" containerID="a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.105892 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea"} err="failed to get container status \"a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea\": rpc error: code = NotFound desc = could not find container \"a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea\": container with ID starting with a0f2cba8fa77e70a4bfcebc1b00b9e73f519a315a91e3e4d1f805b1d5c9267ea not found: ID does not exist" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.105911 4676 scope.go:117] "RemoveContainer" containerID="09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04" Jan 24 00:19:21 crc kubenswrapper[4676]: E0124 00:19:21.106134 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04\": container with ID starting with 09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04 not found: ID does not exist" containerID="09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.106153 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04"} err="failed to get container status \"09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04\": rpc error: code = NotFound desc = could not find container \"09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04\": container with ID starting with 09d568c70998bca76872972662abe01d907ca7c4e29651c8d5b4e3e18f5a2f04 not found: ID does not exist" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.106165 4676 scope.go:117] "RemoveContainer" containerID="fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49" Jan 24 00:19:21 crc kubenswrapper[4676]: E0124 00:19:21.106320 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49\": container with ID starting with fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49 not found: ID does not exist" containerID="fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49" Jan 24 00:19:21 crc kubenswrapper[4676]: I0124 00:19:21.106340 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49"} err="failed to get container status \"fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49\": rpc error: code = NotFound desc = could not find container \"fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49\": container with ID starting with fad1a55313f087cfbdd86fc92712061e080059764bba6dea18e37b485e2e6d49 not found: ID does not exist" Jan 24 00:19:22 crc kubenswrapper[4676]: I0124 00:19:22.263811 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" path="/var/lib/kubelet/pods/6fb6d395-75c7-4f58-85cf-a97ad12bb288/volumes" Jan 24 00:19:22 crc kubenswrapper[4676]: I0124 00:19:22.963992 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:19:22 crc kubenswrapper[4676]: I0124 00:19:22.964061 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:19:23 crc kubenswrapper[4676]: I0124 00:19:23.024099 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:19:23 crc kubenswrapper[4676]: I0124 00:19:23.103163 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:19:24 crc kubenswrapper[4676]: I0124 00:19:24.221125 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8rvgb"] Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.070671 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" event={"ID":"df9ab5f0-f577-4303-8045-f960c67a6936","Type":"ContainerStarted","Data":"ad34eaa169d9abea0b138a3ceff2ef435f1c27d3e02a3f2fb45def6ff0c21ca3"} Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.070898 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8rvgb" podUID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerName="registry-server" containerID="cri-o://055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1" gracePeriod=2 Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.071438 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.105987 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" podStartSLOduration=3.057855358 podStartE2EDuration="1m9.105969583s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.803648501 +0000 UTC m=+882.833619502" lastFinishedPulling="2026-01-24 00:19:24.851762706 +0000 UTC m=+948.881733727" observedRunningTime="2026-01-24 00:19:25.104300701 +0000 UTC m=+949.134271712" watchObservedRunningTime="2026-01-24 00:19:25.105969583 +0000 UTC m=+949.135940594" Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.456780 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.592708 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-catalog-content\") pod \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.592802 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sf2s\" (UniqueName: \"kubernetes.io/projected/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-kube-api-access-4sf2s\") pod \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.592911 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-utilities\") pod \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\" (UID: \"dccfed4b-e148-4e09-8c96-f5a28fefd6b7\") " Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.593936 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-utilities" (OuterVolumeSpecName: "utilities") pod "dccfed4b-e148-4e09-8c96-f5a28fefd6b7" (UID: "dccfed4b-e148-4e09-8c96-f5a28fefd6b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.599170 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-kube-api-access-4sf2s" (OuterVolumeSpecName: "kube-api-access-4sf2s") pod "dccfed4b-e148-4e09-8c96-f5a28fefd6b7" (UID: "dccfed4b-e148-4e09-8c96-f5a28fefd6b7"). InnerVolumeSpecName "kube-api-access-4sf2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.641919 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dccfed4b-e148-4e09-8c96-f5a28fefd6b7" (UID: "dccfed4b-e148-4e09-8c96-f5a28fefd6b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.694254 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.694290 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:19:25 crc kubenswrapper[4676]: I0124 00:19:25.694304 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sf2s\" (UniqueName: \"kubernetes.io/projected/dccfed4b-e148-4e09-8c96-f5a28fefd6b7-kube-api-access-4sf2s\") on node \"crc\" DevicePath \"\"" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.078601 4676 generic.go:334] "Generic (PLEG): container finished" podID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerID="055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1" exitCode=0 Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.078663 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8rvgb" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.078662 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rvgb" event={"ID":"dccfed4b-e148-4e09-8c96-f5a28fefd6b7","Type":"ContainerDied","Data":"055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1"} Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.079006 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8rvgb" event={"ID":"dccfed4b-e148-4e09-8c96-f5a28fefd6b7","Type":"ContainerDied","Data":"ae7b958b56efd4b4607c5292b0d29fa3ac42c0d3302dd61deaf7d117c5fd353a"} Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.079033 4676 scope.go:117] "RemoveContainer" containerID="055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.081534 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" event={"ID":"dd6346d8-9cf1-4364-b480-f4c2d872472f","Type":"ContainerStarted","Data":"766e9849778c46a698401986487d4e96a72a15c35d3b58bfc59ed914aad23d0a"} Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.081924 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.095314 4676 scope.go:117] "RemoveContainer" containerID="d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.124900 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" podStartSLOduration=3.181729177 podStartE2EDuration="1m10.124877894s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.833405548 +0000 UTC m=+882.863376539" lastFinishedPulling="2026-01-24 00:19:25.776554255 +0000 UTC m=+949.806525256" observedRunningTime="2026-01-24 00:19:26.11988738 +0000 UTC m=+950.149858391" watchObservedRunningTime="2026-01-24 00:19:26.124877894 +0000 UTC m=+950.154848895" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.129995 4676 scope.go:117] "RemoveContainer" containerID="8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.137352 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8rvgb"] Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.143343 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8rvgb"] Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.167986 4676 scope.go:117] "RemoveContainer" containerID="055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1" Jan 24 00:19:26 crc kubenswrapper[4676]: E0124 00:19:26.168708 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1\": container with ID starting with 055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1 not found: ID does not exist" containerID="055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.168738 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1"} err="failed to get container status \"055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1\": rpc error: code = NotFound desc = could not find container \"055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1\": container with ID starting with 055fb9a70848ca751070c86e921c80b30bcb2f19ac12925601a33d8c6cc09ff1 not found: ID does not exist" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.168758 4676 scope.go:117] "RemoveContainer" containerID="d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff" Jan 24 00:19:26 crc kubenswrapper[4676]: E0124 00:19:26.169461 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff\": container with ID starting with d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff not found: ID does not exist" containerID="d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.169487 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff"} err="failed to get container status \"d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff\": rpc error: code = NotFound desc = could not find container \"d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff\": container with ID starting with d78da559e910d3a887dafee8ce3b16fd2ae5d6fe6bbaf01fbe69c663c5b3dcff not found: ID does not exist" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.169502 4676 scope.go:117] "RemoveContainer" containerID="8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943" Jan 24 00:19:26 crc kubenswrapper[4676]: E0124 00:19:26.171215 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943\": container with ID starting with 8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943 not found: ID does not exist" containerID="8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.171236 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943"} err="failed to get container status \"8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943\": rpc error: code = NotFound desc = could not find container \"8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943\": container with ID starting with 8860288802f8353ee23e76485e0052d8eb477114812b0d068faa208bb24aa943 not found: ID does not exist" Jan 24 00:19:26 crc kubenswrapper[4676]: I0124 00:19:26.262084 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" path="/var/lib/kubelet/pods/dccfed4b-e148-4e09-8c96-f5a28fefd6b7/volumes" Jan 24 00:19:27 crc kubenswrapper[4676]: I0124 00:19:27.730644 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-gqg82" Jan 24 00:19:29 crc kubenswrapper[4676]: I0124 00:19:29.103531 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" event={"ID":"d85fa79d-818f-4079-aac4-f3fa51a90e9a","Type":"ContainerStarted","Data":"137a6cfdf47ebb609e2b69b5378d7a09fae72d340a63b2433801e97509aeda1c"} Jan 24 00:19:29 crc kubenswrapper[4676]: I0124 00:19:29.104068 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" Jan 24 00:19:29 crc kubenswrapper[4676]: I0124 00:19:29.110539 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" event={"ID":"fce4c8d0-b903-4873-8c89-2f4b9dd9c05d","Type":"ContainerStarted","Data":"9050915a23f402f2f8cd922c2c90d2893c34a808c090b0f7343b1dfffae6d889"} Jan 24 00:19:29 crc kubenswrapper[4676]: I0124 00:19:29.131696 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" podStartSLOduration=3.967131079 podStartE2EDuration="1m13.131678048s" podCreationTimestamp="2026-01-24 00:18:16 +0000 UTC" firstStartedPulling="2026-01-24 00:18:18.853143637 +0000 UTC m=+882.883114638" lastFinishedPulling="2026-01-24 00:19:28.017690606 +0000 UTC m=+952.047661607" observedRunningTime="2026-01-24 00:19:29.1294544 +0000 UTC m=+953.159425411" watchObservedRunningTime="2026-01-24 00:19:29.131678048 +0000 UTC m=+953.161649049" Jan 24 00:19:37 crc kubenswrapper[4676]: I0124 00:19:37.064803 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-wqbcz" Jan 24 00:19:37 crc kubenswrapper[4676]: I0124 00:19:37.080987 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-h44qw" podStartSLOduration=11.213571732 podStartE2EDuration="1m20.08097095s" podCreationTimestamp="2026-01-24 00:18:17 +0000 UTC" firstStartedPulling="2026-01-24 00:18:19.098830241 +0000 UTC m=+883.128801242" lastFinishedPulling="2026-01-24 00:19:27.966229449 +0000 UTC m=+951.996200460" observedRunningTime="2026-01-24 00:19:29.145259246 +0000 UTC m=+953.175230267" watchObservedRunningTime="2026-01-24 00:19:37.08097095 +0000 UTC m=+961.110941951" Jan 24 00:19:37 crc kubenswrapper[4676]: I0124 00:19:37.114928 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6b8bc8d87d-4xp45" Jan 24 00:19:37 crc kubenswrapper[4676]: I0124 00:19:37.585499 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5d646b7d76-tdqwh" Jan 24 00:19:39 crc kubenswrapper[4676]: I0124 00:19:39.364001 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:19:39 crc kubenswrapper[4676]: I0124 00:19:39.364086 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.501967 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tkq64"] Jan 24 00:19:51 crc kubenswrapper[4676]: E0124 00:19:51.513522 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerName="extract-utilities" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.513557 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerName="extract-utilities" Jan 24 00:19:51 crc kubenswrapper[4676]: E0124 00:19:51.513577 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerName="extract-utilities" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.513584 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerName="extract-utilities" Jan 24 00:19:51 crc kubenswrapper[4676]: E0124 00:19:51.513595 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerName="registry-server" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.513611 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerName="registry-server" Jan 24 00:19:51 crc kubenswrapper[4676]: E0124 00:19:51.513630 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerName="extract-content" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.513637 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerName="extract-content" Jan 24 00:19:51 crc kubenswrapper[4676]: E0124 00:19:51.513645 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerName="registry-server" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.513651 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerName="registry-server" Jan 24 00:19:51 crc kubenswrapper[4676]: E0124 00:19:51.513662 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerName="extract-content" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.513669 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerName="extract-content" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.513868 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="dccfed4b-e148-4e09-8c96-f5a28fefd6b7" containerName="registry-server" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.513881 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fb6d395-75c7-4f58-85cf-a97ad12bb288" containerName="registry-server" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.514569 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.518194 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.518419 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-4wchf" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.518571 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.518697 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.529449 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tkq64"] Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.551226 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mj7x9"] Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.552286 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.554467 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.597768 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mj7x9"] Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.615968 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-config\") pod \"dnsmasq-dns-78dd6ddcc-mj7x9\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.616018 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx2sb\" (UniqueName: \"kubernetes.io/projected/51b7a0ef-2050-425e-a0ea-b49b4c39662b-kube-api-access-jx2sb\") pod \"dnsmasq-dns-78dd6ddcc-mj7x9\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.616058 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mj7x9\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.616176 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d55644-d050-4673-9c82-cba0c15d4537-config\") pod \"dnsmasq-dns-675f4bcbfc-tkq64\" (UID: \"67d55644-d050-4673-9c82-cba0c15d4537\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.616238 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npdwx\" (UniqueName: \"kubernetes.io/projected/67d55644-d050-4673-9c82-cba0c15d4537-kube-api-access-npdwx\") pod \"dnsmasq-dns-675f4bcbfc-tkq64\" (UID: \"67d55644-d050-4673-9c82-cba0c15d4537\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.718075 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-config\") pod \"dnsmasq-dns-78dd6ddcc-mj7x9\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.719035 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx2sb\" (UniqueName: \"kubernetes.io/projected/51b7a0ef-2050-425e-a0ea-b49b4c39662b-kube-api-access-jx2sb\") pod \"dnsmasq-dns-78dd6ddcc-mj7x9\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.719421 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mj7x9\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.720077 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d55644-d050-4673-9c82-cba0c15d4537-config\") pod \"dnsmasq-dns-675f4bcbfc-tkq64\" (UID: \"67d55644-d050-4673-9c82-cba0c15d4537\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.720164 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npdwx\" (UniqueName: \"kubernetes.io/projected/67d55644-d050-4673-9c82-cba0c15d4537-kube-api-access-npdwx\") pod \"dnsmasq-dns-675f4bcbfc-tkq64\" (UID: \"67d55644-d050-4673-9c82-cba0c15d4537\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.718992 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-config\") pod \"dnsmasq-dns-78dd6ddcc-mj7x9\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.720047 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mj7x9\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.723250 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d55644-d050-4673-9c82-cba0c15d4537-config\") pod \"dnsmasq-dns-675f4bcbfc-tkq64\" (UID: \"67d55644-d050-4673-9c82-cba0c15d4537\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.738294 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx2sb\" (UniqueName: \"kubernetes.io/projected/51b7a0ef-2050-425e-a0ea-b49b4c39662b-kube-api-access-jx2sb\") pod \"dnsmasq-dns-78dd6ddcc-mj7x9\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.738908 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npdwx\" (UniqueName: \"kubernetes.io/projected/67d55644-d050-4673-9c82-cba0c15d4537-kube-api-access-npdwx\") pod \"dnsmasq-dns-675f4bcbfc-tkq64\" (UID: \"67d55644-d050-4673-9c82-cba0c15d4537\") " pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.828304 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:19:51 crc kubenswrapper[4676]: I0124 00:19:51.866184 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:19:52 crc kubenswrapper[4676]: I0124 00:19:52.302597 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mj7x9"] Jan 24 00:19:52 crc kubenswrapper[4676]: W0124 00:19:52.305016 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51b7a0ef_2050_425e_a0ea_b49b4c39662b.slice/crio-3bc62abc87bbd103f1233186e6fef1bec32233b081b4c253c02f750c9b801731 WatchSource:0}: Error finding container 3bc62abc87bbd103f1233186e6fef1bec32233b081b4c253c02f750c9b801731: Status 404 returned error can't find the container with id 3bc62abc87bbd103f1233186e6fef1bec32233b081b4c253c02f750c9b801731 Jan 24 00:19:52 crc kubenswrapper[4676]: I0124 00:19:52.366081 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tkq64"] Jan 24 00:19:53 crc kubenswrapper[4676]: I0124 00:19:53.279647 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" event={"ID":"67d55644-d050-4673-9c82-cba0c15d4537","Type":"ContainerStarted","Data":"5eee3dc616db8987bcc5c04e8d439d4d666a187be99e08b6950bca9b712da938"} Jan 24 00:19:53 crc kubenswrapper[4676]: I0124 00:19:53.281812 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" event={"ID":"51b7a0ef-2050-425e-a0ea-b49b4c39662b","Type":"ContainerStarted","Data":"3bc62abc87bbd103f1233186e6fef1bec32233b081b4c253c02f750c9b801731"} Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.251274 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tkq64"] Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.290181 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9rgdb"] Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.291200 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.318014 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9rgdb"] Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.372342 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9rgdb\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.372555 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-config\") pod \"dnsmasq-dns-666b6646f7-9rgdb\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.372611 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhwbc\" (UniqueName: \"kubernetes.io/projected/b577ffd6-6d01-45d7-98e5-8c21ac8de280-kube-api-access-hhwbc\") pod \"dnsmasq-dns-666b6646f7-9rgdb\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.474527 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-config\") pod \"dnsmasq-dns-666b6646f7-9rgdb\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.474581 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhwbc\" (UniqueName: \"kubernetes.io/projected/b577ffd6-6d01-45d7-98e5-8c21ac8de280-kube-api-access-hhwbc\") pod \"dnsmasq-dns-666b6646f7-9rgdb\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.474633 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9rgdb\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.475620 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-config\") pod \"dnsmasq-dns-666b6646f7-9rgdb\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.475672 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-dns-svc\") pod \"dnsmasq-dns-666b6646f7-9rgdb\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.521448 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhwbc\" (UniqueName: \"kubernetes.io/projected/b577ffd6-6d01-45d7-98e5-8c21ac8de280-kube-api-access-hhwbc\") pod \"dnsmasq-dns-666b6646f7-9rgdb\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.619856 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.623245 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mj7x9"] Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.657180 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zhmps"] Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.658608 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.735582 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zhmps"] Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.782473 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zhmps\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.782517 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zmnp\" (UniqueName: \"kubernetes.io/projected/b500aa14-7f1d-4209-b0b7-619a37f58bcd-kube-api-access-4zmnp\") pod \"dnsmasq-dns-57d769cc4f-zhmps\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.782546 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-config\") pod \"dnsmasq-dns-57d769cc4f-zhmps\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.884389 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zhmps\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.884440 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zmnp\" (UniqueName: \"kubernetes.io/projected/b500aa14-7f1d-4209-b0b7-619a37f58bcd-kube-api-access-4zmnp\") pod \"dnsmasq-dns-57d769cc4f-zhmps\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.884474 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-config\") pod \"dnsmasq-dns-57d769cc4f-zhmps\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.885303 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-config\") pod \"dnsmasq-dns-57d769cc4f-zhmps\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.888245 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zhmps\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:54 crc kubenswrapper[4676]: I0124 00:19:54.905281 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zmnp\" (UniqueName: \"kubernetes.io/projected/b500aa14-7f1d-4209-b0b7-619a37f58bcd-kube-api-access-4zmnp\") pod \"dnsmasq-dns-57d769cc4f-zhmps\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.036255 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.237626 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9rgdb"] Jan 24 00:19:55 crc kubenswrapper[4676]: W0124 00:19:55.253569 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb577ffd6_6d01_45d7_98e5_8c21ac8de280.slice/crio-f87d0959391165a8c4c5cf866d530b4896a47324d10a6af28b679062fde837bb WatchSource:0}: Error finding container f87d0959391165a8c4c5cf866d530b4896a47324d10a6af28b679062fde837bb: Status 404 returned error can't find the container with id f87d0959391165a8c4c5cf866d530b4896a47324d10a6af28b679062fde837bb Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.303198 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" event={"ID":"b577ffd6-6d01-45d7-98e5-8c21ac8de280","Type":"ContainerStarted","Data":"f87d0959391165a8c4c5cf866d530b4896a47324d10a6af28b679062fde837bb"} Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.454066 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.455543 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.464968 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.465163 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.465288 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nxmk6" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.465426 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.465989 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.466099 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.466221 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.476453 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499444 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499506 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499539 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/68d6466c-a6ff-40ba-952d-007b14efdfd3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499561 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v959\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-kube-api-access-7v959\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499590 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-config-data\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499612 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499817 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499844 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499870 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499902 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.499926 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/68d6466c-a6ff-40ba-952d-007b14efdfd3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601101 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601145 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/68d6466c-a6ff-40ba-952d-007b14efdfd3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601186 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601210 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601234 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/68d6466c-a6ff-40ba-952d-007b14efdfd3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601250 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v959\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-kube-api-access-7v959\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601275 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-config-data\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601309 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601356 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601388 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601415 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.601713 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.603549 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.604117 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.604360 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.605156 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-config-data\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.609527 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.614198 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/68d6466c-a6ff-40ba-952d-007b14efdfd3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.614485 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.653243 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.653334 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/68d6466c-a6ff-40ba-952d-007b14efdfd3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.668173 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v959\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-kube-api-access-7v959\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.682266 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zhmps"] Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.693131 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.800985 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.804084 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.808392 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.813658 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.813832 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.814004 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.814111 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.815993 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.816118 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.817248 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7x4sc" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.820503 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909505 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909565 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909587 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36558de2-6aac-43e9-832d-2f96c46e8152-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909617 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909655 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909736 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36558de2-6aac-43e9-832d-2f96c46e8152-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909767 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909794 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909826 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909873 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdctt\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-kube-api-access-sdctt\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:55 crc kubenswrapper[4676]: I0124 00:19:55.909904 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.010985 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011265 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36558de2-6aac-43e9-832d-2f96c46e8152-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011296 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011341 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011391 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36558de2-6aac-43e9-832d-2f96c46e8152-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011410 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011455 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011474 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011498 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdctt\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-kube-api-access-sdctt\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011517 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.011554 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.012007 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.012764 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.014615 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.015060 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.015699 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.016355 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.017962 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.027993 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36558de2-6aac-43e9-832d-2f96c46e8152-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.034101 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.035870 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdctt\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-kube-api-access-sdctt\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.036229 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36558de2-6aac-43e9-832d-2f96c46e8152-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.056127 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.127243 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.325908 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" event={"ID":"b500aa14-7f1d-4209-b0b7-619a37f58bcd","Type":"ContainerStarted","Data":"98e1e2ee667220310f02f8a03eb3c9bedd6e3875fe7c7233b27e507391445c5f"} Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.347352 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 24 00:19:56 crc kubenswrapper[4676]: W0124 00:19:56.383310 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68d6466c_a6ff_40ba_952d_007b14efdfd3.slice/crio-32030fde9c3a43091d81748f9987539fcf3f4b60bda9f5dc9e85b5f47719d7eb WatchSource:0}: Error finding container 32030fde9c3a43091d81748f9987539fcf3f4b60bda9f5dc9e85b5f47719d7eb: Status 404 returned error can't find the container with id 32030fde9c3a43091d81748f9987539fcf3f4b60bda9f5dc9e85b5f47719d7eb Jan 24 00:19:56 crc kubenswrapper[4676]: I0124 00:19:56.618681 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 24 00:19:56 crc kubenswrapper[4676]: W0124 00:19:56.633805 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36558de2_6aac_43e9_832d_2f96c46e8152.slice/crio-a90a5dbbd754f4d91197a2d8faa756ac77a0f539415213117ebe9918b73147bc WatchSource:0}: Error finding container a90a5dbbd754f4d91197a2d8faa756ac77a0f539415213117ebe9918b73147bc: Status 404 returned error can't find the container with id a90a5dbbd754f4d91197a2d8faa756ac77a0f539415213117ebe9918b73147bc Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.092332 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.093630 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.112454 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.112887 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.113033 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-j4llt" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.113146 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.115984 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.154409 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.233901 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2bbaae64-ac2d-43c6-8984-5483f2eb4211-config-data-default\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.233946 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2bbaae64-ac2d-43c6-8984-5483f2eb4211-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.233984 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbaae64-ac2d-43c6-8984-5483f2eb4211-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.234013 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bbaae64-ac2d-43c6-8984-5483f2eb4211-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.234078 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bbaae64-ac2d-43c6-8984-5483f2eb4211-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.234142 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.234280 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2bbaae64-ac2d-43c6-8984-5483f2eb4211-kolla-config\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.234331 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcvcc\" (UniqueName: \"kubernetes.io/projected/2bbaae64-ac2d-43c6-8984-5483f2eb4211-kube-api-access-tcvcc\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.335728 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbaae64-ac2d-43c6-8984-5483f2eb4211-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.335772 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bbaae64-ac2d-43c6-8984-5483f2eb4211-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.335793 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bbaae64-ac2d-43c6-8984-5483f2eb4211-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.335814 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.335860 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2bbaae64-ac2d-43c6-8984-5483f2eb4211-kolla-config\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.335887 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcvcc\" (UniqueName: \"kubernetes.io/projected/2bbaae64-ac2d-43c6-8984-5483f2eb4211-kube-api-access-tcvcc\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.335913 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2bbaae64-ac2d-43c6-8984-5483f2eb4211-config-data-default\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.335931 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2bbaae64-ac2d-43c6-8984-5483f2eb4211-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.336318 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2bbaae64-ac2d-43c6-8984-5483f2eb4211-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.337421 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.337950 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2bbaae64-ac2d-43c6-8984-5483f2eb4211-config-data-default\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.338805 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2bbaae64-ac2d-43c6-8984-5483f2eb4211-kolla-config\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.339599 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2bbaae64-ac2d-43c6-8984-5483f2eb4211-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.366270 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bbaae64-ac2d-43c6-8984-5483f2eb4211-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.367574 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bbaae64-ac2d-43c6-8984-5483f2eb4211-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.369955 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcvcc\" (UniqueName: \"kubernetes.io/projected/2bbaae64-ac2d-43c6-8984-5483f2eb4211-kube-api-access-tcvcc\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.374573 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"2bbaae64-ac2d-43c6-8984-5483f2eb4211\") " pod="openstack/openstack-galera-0" Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.399835 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"68d6466c-a6ff-40ba-952d-007b14efdfd3","Type":"ContainerStarted","Data":"32030fde9c3a43091d81748f9987539fcf3f4b60bda9f5dc9e85b5f47719d7eb"} Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.420240 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"36558de2-6aac-43e9-832d-2f96c46e8152","Type":"ContainerStarted","Data":"a90a5dbbd754f4d91197a2d8faa756ac77a0f539415213117ebe9918b73147bc"} Jan 24 00:19:57 crc kubenswrapper[4676]: I0124 00:19:57.431726 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.068925 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.309181 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.312572 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.312727 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.315744 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.317718 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-5czjs" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.317812 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.317972 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.458239 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/19365292-50d8-4e94-952f-2df7ee20f0ba-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.458311 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.458352 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19365292-50d8-4e94-952f-2df7ee20f0ba-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.458472 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19365292-50d8-4e94-952f-2df7ee20f0ba-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.458522 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/19365292-50d8-4e94-952f-2df7ee20f0ba-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.458552 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/19365292-50d8-4e94-952f-2df7ee20f0ba-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.458573 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/19365292-50d8-4e94-952f-2df7ee20f0ba-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.458611 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfb6c\" (UniqueName: \"kubernetes.io/projected/19365292-50d8-4e94-952f-2df7ee20f0ba-kube-api-access-cfb6c\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.513839 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.515601 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.521698 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-x2pp8" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.525364 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.525646 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.541254 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.559482 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/19365292-50d8-4e94-952f-2df7ee20f0ba-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.559536 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/19365292-50d8-4e94-952f-2df7ee20f0ba-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.559576 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfb6c\" (UniqueName: \"kubernetes.io/projected/19365292-50d8-4e94-952f-2df7ee20f0ba-kube-api-access-cfb6c\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.559613 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/19365292-50d8-4e94-952f-2df7ee20f0ba-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.559647 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.559676 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19365292-50d8-4e94-952f-2df7ee20f0ba-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.559697 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19365292-50d8-4e94-952f-2df7ee20f0ba-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.559734 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/19365292-50d8-4e94-952f-2df7ee20f0ba-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.560345 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/19365292-50d8-4e94-952f-2df7ee20f0ba-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.560432 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.560651 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/19365292-50d8-4e94-952f-2df7ee20f0ba-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.560730 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/19365292-50d8-4e94-952f-2df7ee20f0ba-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.564538 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19365292-50d8-4e94-952f-2df7ee20f0ba-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.580698 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/19365292-50d8-4e94-952f-2df7ee20f0ba-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.584731 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19365292-50d8-4e94-952f-2df7ee20f0ba-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.591021 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfb6c\" (UniqueName: \"kubernetes.io/projected/19365292-50d8-4e94-952f-2df7ee20f0ba-kube-api-access-cfb6c\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.604852 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"19365292-50d8-4e94-952f-2df7ee20f0ba\") " pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.642980 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.661798 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/545f6045-cf2f-4d4b-91d8-227148ddd71a-config-data\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.661844 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j59wl\" (UniqueName: \"kubernetes.io/projected/545f6045-cf2f-4d4b-91d8-227148ddd71a-kube-api-access-j59wl\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.661872 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/545f6045-cf2f-4d4b-91d8-227148ddd71a-kolla-config\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.661897 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545f6045-cf2f-4d4b-91d8-227148ddd71a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.662247 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/545f6045-cf2f-4d4b-91d8-227148ddd71a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.763874 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/545f6045-cf2f-4d4b-91d8-227148ddd71a-config-data\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.763906 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j59wl\" (UniqueName: \"kubernetes.io/projected/545f6045-cf2f-4d4b-91d8-227148ddd71a-kube-api-access-j59wl\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.763940 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/545f6045-cf2f-4d4b-91d8-227148ddd71a-kolla-config\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.763964 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545f6045-cf2f-4d4b-91d8-227148ddd71a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.765276 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/545f6045-cf2f-4d4b-91d8-227148ddd71a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.768139 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/545f6045-cf2f-4d4b-91d8-227148ddd71a-config-data\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.769131 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/545f6045-cf2f-4d4b-91d8-227148ddd71a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.769673 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/545f6045-cf2f-4d4b-91d8-227148ddd71a-kolla-config\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.789419 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545f6045-cf2f-4d4b-91d8-227148ddd71a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.794562 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j59wl\" (UniqueName: \"kubernetes.io/projected/545f6045-cf2f-4d4b-91d8-227148ddd71a-kube-api-access-j59wl\") pod \"memcached-0\" (UID: \"545f6045-cf2f-4d4b-91d8-227148ddd71a\") " pod="openstack/memcached-0" Jan 24 00:19:58 crc kubenswrapper[4676]: I0124 00:19:58.835968 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 24 00:20:00 crc kubenswrapper[4676]: I0124 00:20:00.339865 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 24 00:20:00 crc kubenswrapper[4676]: I0124 00:20:00.343102 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 24 00:20:00 crc kubenswrapper[4676]: I0124 00:20:00.346828 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-bcjhj" Jan 24 00:20:00 crc kubenswrapper[4676]: I0124 00:20:00.366867 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 24 00:20:00 crc kubenswrapper[4676]: I0124 00:20:00.460127 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqg2d\" (UniqueName: \"kubernetes.io/projected/c626d726-d406-41c7-8e66-07cc148d3aa7-kube-api-access-vqg2d\") pod \"kube-state-metrics-0\" (UID: \"c626d726-d406-41c7-8e66-07cc148d3aa7\") " pod="openstack/kube-state-metrics-0" Jan 24 00:20:00 crc kubenswrapper[4676]: I0124 00:20:00.561736 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqg2d\" (UniqueName: \"kubernetes.io/projected/c626d726-d406-41c7-8e66-07cc148d3aa7-kube-api-access-vqg2d\") pod \"kube-state-metrics-0\" (UID: \"c626d726-d406-41c7-8e66-07cc148d3aa7\") " pod="openstack/kube-state-metrics-0" Jan 24 00:20:00 crc kubenswrapper[4676]: I0124 00:20:00.597628 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqg2d\" (UniqueName: \"kubernetes.io/projected/c626d726-d406-41c7-8e66-07cc148d3aa7-kube-api-access-vqg2d\") pod \"kube-state-metrics-0\" (UID: \"c626d726-d406-41c7-8e66-07cc148d3aa7\") " pod="openstack/kube-state-metrics-0" Jan 24 00:20:00 crc kubenswrapper[4676]: I0124 00:20:00.672314 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.237190 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kvzq7"] Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.239302 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.246925 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kvzq7"] Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.289161 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-catalog-content\") pod \"certified-operators-kvzq7\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.289227 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-utilities\") pod \"certified-operators-kvzq7\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.289277 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk82l\" (UniqueName: \"kubernetes.io/projected/3df4276b-01b4-403f-b040-85d5a8a9ef03-kube-api-access-tk82l\") pod \"certified-operators-kvzq7\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.390543 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-catalog-content\") pod \"certified-operators-kvzq7\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.390613 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-utilities\") pod \"certified-operators-kvzq7\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.390665 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk82l\" (UniqueName: \"kubernetes.io/projected/3df4276b-01b4-403f-b040-85d5a8a9ef03-kube-api-access-tk82l\") pod \"certified-operators-kvzq7\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.391168 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-catalog-content\") pod \"certified-operators-kvzq7\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.391226 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-utilities\") pod \"certified-operators-kvzq7\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.408993 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk82l\" (UniqueName: \"kubernetes.io/projected/3df4276b-01b4-403f-b040-85d5a8a9ef03-kube-api-access-tk82l\") pod \"certified-operators-kvzq7\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:02 crc kubenswrapper[4676]: I0124 00:20:02.567290 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.018794 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jsd4n"] Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.020703 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.025133 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-brqw8" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.025217 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.026287 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.038478 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jsd4n"] Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.084012 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-6sl9q"] Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.088644 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.091394 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6sl9q"] Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.140429 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7adbcf83-efbd-4e8d-97e5-f8768463284a-scripts\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.140490 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8blx\" (UniqueName: \"kubernetes.io/projected/7adbcf83-efbd-4e8d-97e5-f8768463284a-kube-api-access-m8blx\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.140559 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7adbcf83-efbd-4e8d-97e5-f8768463284a-combined-ca-bundle\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.140649 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7adbcf83-efbd-4e8d-97e5-f8768463284a-ovn-controller-tls-certs\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.140703 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7adbcf83-efbd-4e8d-97e5-f8768463284a-var-log-ovn\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.140759 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7adbcf83-efbd-4e8d-97e5-f8768463284a-var-run-ovn\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.140795 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7adbcf83-efbd-4e8d-97e5-f8768463284a-var-run\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242006 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-var-lib\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242048 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7adbcf83-efbd-4e8d-97e5-f8768463284a-combined-ca-bundle\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242098 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7adbcf83-efbd-4e8d-97e5-f8768463284a-ovn-controller-tls-certs\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242117 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/427fbd2d-16ef-44a6-a71d-8172f56b863d-scripts\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242147 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-etc-ovs\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242174 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7adbcf83-efbd-4e8d-97e5-f8768463284a-var-log-ovn\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242201 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7adbcf83-efbd-4e8d-97e5-f8768463284a-var-run-ovn\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242226 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7adbcf83-efbd-4e8d-97e5-f8768463284a-var-run\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242249 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpbnv\" (UniqueName: \"kubernetes.io/projected/427fbd2d-16ef-44a6-a71d-8172f56b863d-kube-api-access-fpbnv\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242269 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-var-run\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242287 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7adbcf83-efbd-4e8d-97e5-f8768463284a-scripts\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242311 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-var-log\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.242332 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8blx\" (UniqueName: \"kubernetes.io/projected/7adbcf83-efbd-4e8d-97e5-f8768463284a-kube-api-access-m8blx\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.243764 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7adbcf83-efbd-4e8d-97e5-f8768463284a-var-run-ovn\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.243837 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7adbcf83-efbd-4e8d-97e5-f8768463284a-var-log-ovn\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.245346 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7adbcf83-efbd-4e8d-97e5-f8768463284a-scripts\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.245492 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7adbcf83-efbd-4e8d-97e5-f8768463284a-var-run\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.247318 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/7adbcf83-efbd-4e8d-97e5-f8768463284a-ovn-controller-tls-certs\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.258647 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7adbcf83-efbd-4e8d-97e5-f8768463284a-combined-ca-bundle\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.271958 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8blx\" (UniqueName: \"kubernetes.io/projected/7adbcf83-efbd-4e8d-97e5-f8768463284a-kube-api-access-m8blx\") pod \"ovn-controller-jsd4n\" (UID: \"7adbcf83-efbd-4e8d-97e5-f8768463284a\") " pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.337815 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.343665 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/427fbd2d-16ef-44a6-a71d-8172f56b863d-scripts\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.343820 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-etc-ovs\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.343971 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpbnv\" (UniqueName: \"kubernetes.io/projected/427fbd2d-16ef-44a6-a71d-8172f56b863d-kube-api-access-fpbnv\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.344076 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-var-run\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.344190 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-var-log\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.344299 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-var-lib\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.344081 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-etc-ovs\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.344137 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-var-run\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.344345 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-var-log\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.344610 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/427fbd2d-16ef-44a6-a71d-8172f56b863d-var-lib\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.346184 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/427fbd2d-16ef-44a6-a71d-8172f56b863d-scripts\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.362414 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpbnv\" (UniqueName: \"kubernetes.io/projected/427fbd2d-16ef-44a6-a71d-8172f56b863d-kube-api-access-fpbnv\") pod \"ovn-controller-ovs-6sl9q\" (UID: \"427fbd2d-16ef-44a6-a71d-8172f56b863d\") " pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.402593 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.895815 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.897214 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.900880 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.901066 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-wj2kh" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.901089 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.901204 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.906389 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Jan 24 00:20:05 crc kubenswrapper[4676]: I0124 00:20:05.932885 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.054446 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3bff778a-b10f-4ba9-a12f-f4086608fd30-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.054512 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bff778a-b10f-4ba9-a12f-f4086608fd30-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.054554 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.054619 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3bff778a-b10f-4ba9-a12f-f4086608fd30-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.054684 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bff778a-b10f-4ba9-a12f-f4086608fd30-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.054737 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bff778a-b10f-4ba9-a12f-f4086608fd30-config\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.054782 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bff778a-b10f-4ba9-a12f-f4086608fd30-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.054816 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tsl7\" (UniqueName: \"kubernetes.io/projected/3bff778a-b10f-4ba9-a12f-f4086608fd30-kube-api-access-6tsl7\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.156481 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3bff778a-b10f-4ba9-a12f-f4086608fd30-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.156554 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bff778a-b10f-4ba9-a12f-f4086608fd30-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.156597 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.156662 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3bff778a-b10f-4ba9-a12f-f4086608fd30-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.156755 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bff778a-b10f-4ba9-a12f-f4086608fd30-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.156816 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bff778a-b10f-4ba9-a12f-f4086608fd30-config\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.156872 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bff778a-b10f-4ba9-a12f-f4086608fd30-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.156905 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tsl7\" (UniqueName: \"kubernetes.io/projected/3bff778a-b10f-4ba9-a12f-f4086608fd30-kube-api-access-6tsl7\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.157099 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.157551 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3bff778a-b10f-4ba9-a12f-f4086608fd30-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.158621 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bff778a-b10f-4ba9-a12f-f4086608fd30-config\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.162010 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3bff778a-b10f-4ba9-a12f-f4086608fd30-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.169526 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bff778a-b10f-4ba9-a12f-f4086608fd30-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.169742 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bff778a-b10f-4ba9-a12f-f4086608fd30-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.169844 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bff778a-b10f-4ba9-a12f-f4086608fd30-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.174857 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tsl7\" (UniqueName: \"kubernetes.io/projected/3bff778a-b10f-4ba9-a12f-f4086608fd30-kube-api-access-6tsl7\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.185870 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-nb-0\" (UID: \"3bff778a-b10f-4ba9-a12f-f4086608fd30\") " pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:06 crc kubenswrapper[4676]: I0124 00:20:06.222900 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.726615 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.728682 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.731310 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.732823 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.733346 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vwpk5" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.735860 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.743924 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.880579 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.880684 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-config\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.880766 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.880898 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.880957 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.881002 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.881091 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.881154 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrj7g\" (UniqueName: \"kubernetes.io/projected/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-kube-api-access-zrj7g\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.983005 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.983466 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrj7g\" (UniqueName: \"kubernetes.io/projected/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-kube-api-access-zrj7g\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.984012 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.984173 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-config\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.985154 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.985293 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.985793 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.986250 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.985194 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.985093 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-config\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.985749 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.986655 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.990848 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:07 crc kubenswrapper[4676]: I0124 00:20:07.992078 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:08 crc kubenswrapper[4676]: I0124 00:20:07.999993 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:08 crc kubenswrapper[4676]: I0124 00:20:08.008196 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrj7g\" (UniqueName: \"kubernetes.io/projected/f3648c65-9fcb-4a9e-b4cb-d8437dc00141-kube-api-access-zrj7g\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:08 crc kubenswrapper[4676]: I0124 00:20:08.013812 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f3648c65-9fcb-4a9e-b4cb-d8437dc00141\") " pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:08 crc kubenswrapper[4676]: I0124 00:20:08.047023 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:09 crc kubenswrapper[4676]: I0124 00:20:09.363989 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:20:09 crc kubenswrapper[4676]: I0124 00:20:09.364071 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:20:09 crc kubenswrapper[4676]: I0124 00:20:09.364129 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:20:09 crc kubenswrapper[4676]: I0124 00:20:09.365059 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2687110039e3aba350a72ed3647bbafb008d22f301a8b50baa7159c6eca5ba33"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:20:09 crc kubenswrapper[4676]: I0124 00:20:09.365144 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://2687110039e3aba350a72ed3647bbafb008d22f301a8b50baa7159c6eca5ba33" gracePeriod=600 Jan 24 00:20:09 crc kubenswrapper[4676]: I0124 00:20:09.568723 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="2687110039e3aba350a72ed3647bbafb008d22f301a8b50baa7159c6eca5ba33" exitCode=0 Jan 24 00:20:09 crc kubenswrapper[4676]: I0124 00:20:09.568792 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"2687110039e3aba350a72ed3647bbafb008d22f301a8b50baa7159c6eca5ba33"} Jan 24 00:20:09 crc kubenswrapper[4676]: I0124 00:20:09.569089 4676 scope.go:117] "RemoveContainer" containerID="daf4dfd81dc7faee8c5a37cce872ffde5731f2d91708788dd42d2993fec18ba6" Jan 24 00:20:11 crc kubenswrapper[4676]: I0124 00:20:11.588731 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2bbaae64-ac2d-43c6-8984-5483f2eb4211","Type":"ContainerStarted","Data":"a1c59ca6b94252368f835fcfdca3061486fe13cfd8944f4df5139db16f75750c"} Jan 24 00:20:14 crc kubenswrapper[4676]: E0124 00:20:14.660205 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Jan 24 00:20:14 crc kubenswrapper[4676]: E0124 00:20:14.660661 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7v959,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(68d6466c-a6ff-40ba-952d-007b14efdfd3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:20:14 crc kubenswrapper[4676]: E0124 00:20:14.661948 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" Jan 24 00:20:14 crc kubenswrapper[4676]: E0124 00:20:14.690171 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Jan 24 00:20:14 crc kubenswrapper[4676]: E0124 00:20:14.690558 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sdctt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(36558de2-6aac-43e9-832d-2f96c46e8152): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:20:14 crc kubenswrapper[4676]: E0124 00:20:14.692228 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="36558de2-6aac-43e9-832d-2f96c46e8152" Jan 24 00:20:15 crc kubenswrapper[4676]: I0124 00:20:15.037013 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 24 00:20:15 crc kubenswrapper[4676]: E0124 00:20:15.624783 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="36558de2-6aac-43e9-832d-2f96c46e8152" Jan 24 00:20:15 crc kubenswrapper[4676]: E0124 00:20:15.630810 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" Jan 24 00:20:21 crc kubenswrapper[4676]: I0124 00:20:21.674661 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"545f6045-cf2f-4d4b-91d8-227148ddd71a","Type":"ContainerStarted","Data":"129868f7f45e8628b78e5f212f82e735051410c9a3c82315b54413272f8815b7"} Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.868401 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.868588 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jx2sb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-mj7x9_openstack(51b7a0ef-2050-425e-a0ea-b49b4c39662b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.869806 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" podUID="51b7a0ef-2050-425e-a0ea-b49b4c39662b" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.881348 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.881500 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4zmnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-zhmps_openstack(b500aa14-7f1d-4209-b0b7-619a37f58bcd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.883593 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" podUID="b500aa14-7f1d-4209-b0b7-619a37f58bcd" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.908644 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.908783 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hhwbc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-9rgdb_openstack(b577ffd6-6d01-45d7-98e5-8c21ac8de280): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.909966 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" podUID="b577ffd6-6d01-45d7-98e5-8c21ac8de280" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.922110 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.922533 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-npdwx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-tkq64_openstack(67d55644-d050-4673-9c82-cba0c15d4537): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:20:22 crc kubenswrapper[4676]: E0124 00:20:22.924238 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" podUID="67d55644-d050-4673-9c82-cba0c15d4537" Jan 24 00:20:23 crc kubenswrapper[4676]: I0124 00:20:23.434280 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 24 00:20:23 crc kubenswrapper[4676]: E0124 00:20:23.700414 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" podUID="b500aa14-7f1d-4209-b0b7-619a37f58bcd" Jan 24 00:20:23 crc kubenswrapper[4676]: E0124 00:20:23.700536 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" podUID="b577ffd6-6d01-45d7-98e5-8c21ac8de280" Jan 24 00:20:25 crc kubenswrapper[4676]: E0124 00:20:25.182458 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Jan 24 00:20:25 crc kubenswrapper[4676]: E0124 00:20:25.182907 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tcvcc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(2bbaae64-ac2d-43c6-8984-5483f2eb4211): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:20:25 crc kubenswrapper[4676]: E0124 00:20:25.184435 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="2bbaae64-ac2d-43c6-8984-5483f2eb4211" Jan 24 00:20:25 crc kubenswrapper[4676]: W0124 00:20:25.205661 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3648c65_9fcb_4a9e_b4cb_d8437dc00141.slice/crio-1a8a42fc79fdf52f814acc2ac5d113099eb491ced0b8ac0af600ea3a850f3feb WatchSource:0}: Error finding container 1a8a42fc79fdf52f814acc2ac5d113099eb491ced0b8ac0af600ea3a850f3feb: Status 404 returned error can't find the container with id 1a8a42fc79fdf52f814acc2ac5d113099eb491ced0b8ac0af600ea3a850f3feb Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.304971 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.368120 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.397306 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d55644-d050-4673-9c82-cba0c15d4537-config\") pod \"67d55644-d050-4673-9c82-cba0c15d4537\" (UID: \"67d55644-d050-4673-9c82-cba0c15d4537\") " Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.397347 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npdwx\" (UniqueName: \"kubernetes.io/projected/67d55644-d050-4673-9c82-cba0c15d4537-kube-api-access-npdwx\") pod \"67d55644-d050-4673-9c82-cba0c15d4537\" (UID: \"67d55644-d050-4673-9c82-cba0c15d4537\") " Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.401498 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67d55644-d050-4673-9c82-cba0c15d4537-config" (OuterVolumeSpecName: "config") pod "67d55644-d050-4673-9c82-cba0c15d4537" (UID: "67d55644-d050-4673-9c82-cba0c15d4537"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.411097 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d55644-d050-4673-9c82-cba0c15d4537-kube-api-access-npdwx" (OuterVolumeSpecName: "kube-api-access-npdwx") pod "67d55644-d050-4673-9c82-cba0c15d4537" (UID: "67d55644-d050-4673-9c82-cba0c15d4537"). InnerVolumeSpecName "kube-api-access-npdwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.501969 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-config\") pod \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.502073 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx2sb\" (UniqueName: \"kubernetes.io/projected/51b7a0ef-2050-425e-a0ea-b49b4c39662b-kube-api-access-jx2sb\") pod \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.502115 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-dns-svc\") pod \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\" (UID: \"51b7a0ef-2050-425e-a0ea-b49b4c39662b\") " Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.502400 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67d55644-d050-4673-9c82-cba0c15d4537-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.502417 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npdwx\" (UniqueName: \"kubernetes.io/projected/67d55644-d050-4673-9c82-cba0c15d4537-kube-api-access-npdwx\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.502703 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-config" (OuterVolumeSpecName: "config") pod "51b7a0ef-2050-425e-a0ea-b49b4c39662b" (UID: "51b7a0ef-2050-425e-a0ea-b49b4c39662b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.502884 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51b7a0ef-2050-425e-a0ea-b49b4c39662b" (UID: "51b7a0ef-2050-425e-a0ea-b49b4c39662b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.505639 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51b7a0ef-2050-425e-a0ea-b49b4c39662b-kube-api-access-jx2sb" (OuterVolumeSpecName: "kube-api-access-jx2sb") pod "51b7a0ef-2050-425e-a0ea-b49b4c39662b" (UID: "51b7a0ef-2050-425e-a0ea-b49b4c39662b"). InnerVolumeSpecName "kube-api-access-jx2sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.604498 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.604532 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx2sb\" (UniqueName: \"kubernetes.io/projected/51b7a0ef-2050-425e-a0ea-b49b4c39662b-kube-api-access-jx2sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.604555 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51b7a0ef-2050-425e-a0ea-b49b4c39662b-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.717412 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" event={"ID":"67d55644-d050-4673-9c82-cba0c15d4537","Type":"ContainerDied","Data":"5eee3dc616db8987bcc5c04e8d439d4d666a187be99e08b6950bca9b712da938"} Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.717425 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-tkq64" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.718904 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" event={"ID":"51b7a0ef-2050-425e-a0ea-b49b4c39662b","Type":"ContainerDied","Data":"3bc62abc87bbd103f1233186e6fef1bec32233b081b4c253c02f750c9b801731"} Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.718964 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mj7x9" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.730912 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f3648c65-9fcb-4a9e-b4cb-d8437dc00141","Type":"ContainerStarted","Data":"1a8a42fc79fdf52f814acc2ac5d113099eb491ced0b8ac0af600ea3a850f3feb"} Jan 24 00:20:25 crc kubenswrapper[4676]: E0124 00:20:25.737500 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="2bbaae64-ac2d-43c6-8984-5483f2eb4211" Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.771733 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.829414 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tkq64"] Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.844008 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-tkq64"] Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.869456 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mj7x9"] Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.874789 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mj7x9"] Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.959150 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.967879 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-6sl9q"] Jan 24 00:20:25 crc kubenswrapper[4676]: I0124 00:20:25.984303 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kvzq7"] Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.030627 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.125329 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jsd4n"] Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.276848 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51b7a0ef-2050-425e-a0ea-b49b4c39662b" path="/var/lib/kubelet/pods/51b7a0ef-2050-425e-a0ea-b49b4c39662b/volumes" Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.281347 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d55644-d050-4673-9c82-cba0c15d4537" path="/var/lib/kubelet/pods/67d55644-d050-4673-9c82-cba0c15d4537/volumes" Jan 24 00:20:26 crc kubenswrapper[4676]: W0124 00:20:26.594266 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3df4276b_01b4_403f_b040_85d5a8a9ef03.slice/crio-88c849f82d3193bf091f531d74d4fde6ab5977e0c495ee6e86efeb15c6348494 WatchSource:0}: Error finding container 88c849f82d3193bf091f531d74d4fde6ab5977e0c495ee6e86efeb15c6348494: Status 404 returned error can't find the container with id 88c849f82d3193bf091f531d74d4fde6ab5977e0c495ee6e86efeb15c6348494 Jan 24 00:20:26 crc kubenswrapper[4676]: W0124 00:20:26.601697 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19365292_50d8_4e94_952f_2df7ee20f0ba.slice/crio-cf388a4e0df4daefa3cbe9e8d89fe266c7758649deb6e360cf7df44e13eba394 WatchSource:0}: Error finding container cf388a4e0df4daefa3cbe9e8d89fe266c7758649deb6e360cf7df44e13eba394: Status 404 returned error can't find the container with id cf388a4e0df4daefa3cbe9e8d89fe266c7758649deb6e360cf7df44e13eba394 Jan 24 00:20:26 crc kubenswrapper[4676]: W0124 00:20:26.603963 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod427fbd2d_16ef_44a6_a71d_8172f56b863d.slice/crio-c93aad98f7fa07e20203ee19cebf86d71af1c830ae464b14a02d327a2041bcf7 WatchSource:0}: Error finding container c93aad98f7fa07e20203ee19cebf86d71af1c830ae464b14a02d327a2041bcf7: Status 404 returned error can't find the container with id c93aad98f7fa07e20203ee19cebf86d71af1c830ae464b14a02d327a2041bcf7 Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.748665 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6sl9q" event={"ID":"427fbd2d-16ef-44a6-a71d-8172f56b863d","Type":"ContainerStarted","Data":"c93aad98f7fa07e20203ee19cebf86d71af1c830ae464b14a02d327a2041bcf7"} Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.752806 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"19365292-50d8-4e94-952f-2df7ee20f0ba","Type":"ContainerStarted","Data":"cf388a4e0df4daefa3cbe9e8d89fe266c7758649deb6e360cf7df44e13eba394"} Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.754531 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c626d726-d406-41c7-8e66-07cc148d3aa7","Type":"ContainerStarted","Data":"a0d93a838adacca6f5b952cbfd2b7bfc9bf0b2ee8a2c8c9cf055df1b2e54dbca"} Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.756120 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jsd4n" event={"ID":"7adbcf83-efbd-4e8d-97e5-f8768463284a","Type":"ContainerStarted","Data":"5e5c0ff40f6a2e48e361d3f8d5e25ed79d68cc33338a1264dfd1615c43c39542"} Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.757252 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvzq7" event={"ID":"3df4276b-01b4-403f-b040-85d5a8a9ef03","Type":"ContainerStarted","Data":"88c849f82d3193bf091f531d74d4fde6ab5977e0c495ee6e86efeb15c6348494"} Jan 24 00:20:26 crc kubenswrapper[4676]: I0124 00:20:26.758520 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3bff778a-b10f-4ba9-a12f-f4086608fd30","Type":"ContainerStarted","Data":"2482cd974ac4c9436d361422f20d218cb4125f3a7bc21fdfea7285523bc780a7"} Jan 24 00:20:27 crc kubenswrapper[4676]: I0124 00:20:27.767843 4676 generic.go:334] "Generic (PLEG): container finished" podID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerID="82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e" exitCode=0 Jan 24 00:20:27 crc kubenswrapper[4676]: I0124 00:20:27.767942 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvzq7" event={"ID":"3df4276b-01b4-403f-b040-85d5a8a9ef03","Type":"ContainerDied","Data":"82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e"} Jan 24 00:20:27 crc kubenswrapper[4676]: I0124 00:20:27.771403 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"545f6045-cf2f-4d4b-91d8-227148ddd71a","Type":"ContainerStarted","Data":"e85e704d14dd52ead197be8059609934dceae01cbb53278d3e8039062faf3104"} Jan 24 00:20:27 crc kubenswrapper[4676]: I0124 00:20:27.771481 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Jan 24 00:20:27 crc kubenswrapper[4676]: I0124 00:20:27.775431 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"f13744ab61f6ff84c30249dfd3e19836649d7bb6c4e4a3db144939c565fd684d"} Jan 24 00:20:27 crc kubenswrapper[4676]: I0124 00:20:27.844434 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=24.634153511 podStartE2EDuration="29.844416332s" podCreationTimestamp="2026-01-24 00:19:58 +0000 UTC" firstStartedPulling="2026-01-24 00:20:21.45065931 +0000 UTC m=+1005.480630331" lastFinishedPulling="2026-01-24 00:20:26.660922141 +0000 UTC m=+1010.690893152" observedRunningTime="2026-01-24 00:20:27.81791752 +0000 UTC m=+1011.847888521" watchObservedRunningTime="2026-01-24 00:20:27.844416332 +0000 UTC m=+1011.874387333" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.848755 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-mq6ns"] Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.850109 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.857591 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.873689 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mq6ns"] Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.893633 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/21550436-cd71-46d8-838e-c51c19ddf8ff-ovs-rundir\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.893704 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/21550436-cd71-46d8-838e-c51c19ddf8ff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.893735 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blhn8\" (UniqueName: \"kubernetes.io/projected/21550436-cd71-46d8-838e-c51c19ddf8ff-kube-api-access-blhn8\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.893801 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21550436-cd71-46d8-838e-c51c19ddf8ff-config\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.893825 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21550436-cd71-46d8-838e-c51c19ddf8ff-combined-ca-bundle\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.893850 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/21550436-cd71-46d8-838e-c51c19ddf8ff-ovn-rundir\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.988724 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zhmps"] Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.994849 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21550436-cd71-46d8-838e-c51c19ddf8ff-config\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.994894 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21550436-cd71-46d8-838e-c51c19ddf8ff-combined-ca-bundle\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.994926 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/21550436-cd71-46d8-838e-c51c19ddf8ff-ovn-rundir\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.994970 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/21550436-cd71-46d8-838e-c51c19ddf8ff-ovs-rundir\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.994990 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/21550436-cd71-46d8-838e-c51c19ddf8ff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.995013 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blhn8\" (UniqueName: \"kubernetes.io/projected/21550436-cd71-46d8-838e-c51c19ddf8ff-kube-api-access-blhn8\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.995874 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21550436-cd71-46d8-838e-c51c19ddf8ff-config\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.999512 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/21550436-cd71-46d8-838e-c51c19ddf8ff-ovs-rundir\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:29 crc kubenswrapper[4676]: I0124 00:20:29.999555 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/21550436-cd71-46d8-838e-c51c19ddf8ff-ovn-rundir\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.015893 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/21550436-cd71-46d8-838e-c51c19ddf8ff-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.017961 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21550436-cd71-46d8-838e-c51c19ddf8ff-combined-ca-bundle\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.032261 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blhn8\" (UniqueName: \"kubernetes.io/projected/21550436-cd71-46d8-838e-c51c19ddf8ff-kube-api-access-blhn8\") pod \"ovn-controller-metrics-mq6ns\" (UID: \"21550436-cd71-46d8-838e-c51c19ddf8ff\") " pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.054425 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qhgmq"] Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.056524 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.058984 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.066908 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qhgmq"] Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.096141 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.096389 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx7ts\" (UniqueName: \"kubernetes.io/projected/88763e6b-93f1-4b9d-bb0c-5c487659691a-kube-api-access-gx7ts\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.096450 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-config\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.096507 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.188213 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mq6ns" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.199152 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.199197 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx7ts\" (UniqueName: \"kubernetes.io/projected/88763e6b-93f1-4b9d-bb0c-5c487659691a-kube-api-access-gx7ts\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.199261 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-config\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.199325 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.200443 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-config\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.200473 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.200502 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.223404 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx7ts\" (UniqueName: \"kubernetes.io/projected/88763e6b-93f1-4b9d-bb0c-5c487659691a-kube-api-access-gx7ts\") pod \"dnsmasq-dns-5bf47b49b7-qhgmq\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.327808 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9rgdb"] Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.366609 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-sbwwk"] Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.367806 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.386295 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.396322 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-sbwwk"] Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.402176 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-dns-svc\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.402238 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.402267 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.402360 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-config\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.402409 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btl5m\" (UniqueName: \"kubernetes.io/projected/1f21d906-2974-4e56-b11f-d888e452c565-kube-api-access-btl5m\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.425388 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.505117 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.505619 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-config\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.506573 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-config\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.506624 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btl5m\" (UniqueName: \"kubernetes.io/projected/1f21d906-2974-4e56-b11f-d888e452c565-kube-api-access-btl5m\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.506717 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-dns-svc\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.506750 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.509952 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.509988 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-dns-svc\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.511478 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.546335 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btl5m\" (UniqueName: \"kubernetes.io/projected/1f21d906-2974-4e56-b11f-d888e452c565-kube-api-access-btl5m\") pod \"dnsmasq-dns-8554648995-sbwwk\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:30 crc kubenswrapper[4676]: I0124 00:20:30.712654 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:31 crc kubenswrapper[4676]: I0124 00:20:31.953083 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:20:31 crc kubenswrapper[4676]: I0124 00:20:31.960667 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.034723 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhwbc\" (UniqueName: \"kubernetes.io/projected/b577ffd6-6d01-45d7-98e5-8c21ac8de280-kube-api-access-hhwbc\") pod \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.034806 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-dns-svc\") pod \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.034913 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-config\") pod \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.034978 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-dns-svc\") pod \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.035097 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zmnp\" (UniqueName: \"kubernetes.io/projected/b500aa14-7f1d-4209-b0b7-619a37f58bcd-kube-api-access-4zmnp\") pod \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\" (UID: \"b500aa14-7f1d-4209-b0b7-619a37f58bcd\") " Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.035146 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-config\") pod \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\" (UID: \"b577ffd6-6d01-45d7-98e5-8c21ac8de280\") " Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.036529 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-config" (OuterVolumeSpecName: "config") pod "b577ffd6-6d01-45d7-98e5-8c21ac8de280" (UID: "b577ffd6-6d01-45d7-98e5-8c21ac8de280"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.037460 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b577ffd6-6d01-45d7-98e5-8c21ac8de280" (UID: "b577ffd6-6d01-45d7-98e5-8c21ac8de280"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.037547 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-config" (OuterVolumeSpecName: "config") pod "b500aa14-7f1d-4209-b0b7-619a37f58bcd" (UID: "b500aa14-7f1d-4209-b0b7-619a37f58bcd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.037913 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b500aa14-7f1d-4209-b0b7-619a37f58bcd" (UID: "b500aa14-7f1d-4209-b0b7-619a37f58bcd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.051528 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b577ffd6-6d01-45d7-98e5-8c21ac8de280-kube-api-access-hhwbc" (OuterVolumeSpecName: "kube-api-access-hhwbc") pod "b577ffd6-6d01-45d7-98e5-8c21ac8de280" (UID: "b577ffd6-6d01-45d7-98e5-8c21ac8de280"). InnerVolumeSpecName "kube-api-access-hhwbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.051608 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b500aa14-7f1d-4209-b0b7-619a37f58bcd-kube-api-access-4zmnp" (OuterVolumeSpecName: "kube-api-access-4zmnp") pod "b500aa14-7f1d-4209-b0b7-619a37f58bcd" (UID: "b500aa14-7f1d-4209-b0b7-619a37f58bcd"). InnerVolumeSpecName "kube-api-access-4zmnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.139060 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhwbc\" (UniqueName: \"kubernetes.io/projected/b577ffd6-6d01-45d7-98e5-8c21ac8de280-kube-api-access-hhwbc\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.139103 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.139115 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.139127 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b500aa14-7f1d-4209-b0b7-619a37f58bcd-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.139141 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zmnp\" (UniqueName: \"kubernetes.io/projected/b500aa14-7f1d-4209-b0b7-619a37f58bcd-kube-api-access-4zmnp\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.139152 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b577ffd6-6d01-45d7-98e5-8c21ac8de280-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.819666 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" event={"ID":"b500aa14-7f1d-4209-b0b7-619a37f58bcd","Type":"ContainerDied","Data":"98e1e2ee667220310f02f8a03eb3c9bedd6e3875fe7c7233b27e507391445c5f"} Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.819718 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zhmps" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.821866 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.823924 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-9rgdb" event={"ID":"b577ffd6-6d01-45d7-98e5-8c21ac8de280","Type":"ContainerDied","Data":"f87d0959391165a8c4c5cf866d530b4896a47324d10a6af28b679062fde837bb"} Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.869055 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9rgdb"] Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.895467 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-9rgdb"] Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.907151 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zhmps"] Jan 24 00:20:32 crc kubenswrapper[4676]: I0124 00:20:32.914414 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zhmps"] Jan 24 00:20:33 crc kubenswrapper[4676]: I0124 00:20:33.832588 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mq6ns"] Jan 24 00:20:33 crc kubenswrapper[4676]: I0124 00:20:33.837574 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Jan 24 00:20:33 crc kubenswrapper[4676]: I0124 00:20:33.973301 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qhgmq"] Jan 24 00:20:34 crc kubenswrapper[4676]: I0124 00:20:34.032816 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-sbwwk"] Jan 24 00:20:34 crc kubenswrapper[4676]: W0124 00:20:34.049412 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88763e6b_93f1_4b9d_bb0c_5c487659691a.slice/crio-e7b4b194ef042a28036832a666cfba7e94711668d0bfc32ab0d45edb956b6341 WatchSource:0}: Error finding container e7b4b194ef042a28036832a666cfba7e94711668d0bfc32ab0d45edb956b6341: Status 404 returned error can't find the container with id e7b4b194ef042a28036832a666cfba7e94711668d0bfc32ab0d45edb956b6341 Jan 24 00:20:34 crc kubenswrapper[4676]: W0124 00:20:34.050240 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f21d906_2974_4e56_b11f_d888e452c565.slice/crio-089f8e4a35ef84364abbbae1879307f5831ba1a07f81a4231c6cfc6207ff2e6a WatchSource:0}: Error finding container 089f8e4a35ef84364abbbae1879307f5831ba1a07f81a4231c6cfc6207ff2e6a: Status 404 returned error can't find the container with id 089f8e4a35ef84364abbbae1879307f5831ba1a07f81a4231c6cfc6207ff2e6a Jan 24 00:20:34 crc kubenswrapper[4676]: I0124 00:20:34.266917 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b500aa14-7f1d-4209-b0b7-619a37f58bcd" path="/var/lib/kubelet/pods/b500aa14-7f1d-4209-b0b7-619a37f58bcd/volumes" Jan 24 00:20:34 crc kubenswrapper[4676]: I0124 00:20:34.267347 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b577ffd6-6d01-45d7-98e5-8c21ac8de280" path="/var/lib/kubelet/pods/b577ffd6-6d01-45d7-98e5-8c21ac8de280/volumes" Jan 24 00:20:34 crc kubenswrapper[4676]: I0124 00:20:34.848770 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-sbwwk" event={"ID":"1f21d906-2974-4e56-b11f-d888e452c565","Type":"ContainerStarted","Data":"089f8e4a35ef84364abbbae1879307f5831ba1a07f81a4231c6cfc6207ff2e6a"} Jan 24 00:20:34 crc kubenswrapper[4676]: I0124 00:20:34.852503 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mq6ns" event={"ID":"21550436-cd71-46d8-838e-c51c19ddf8ff","Type":"ContainerStarted","Data":"bb815329b234bedc0ff99ce46c1b957b19b8f16ce6d1ed4c91a62db36d7a60a6"} Jan 24 00:20:34 crc kubenswrapper[4676]: I0124 00:20:34.854735 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"19365292-50d8-4e94-952f-2df7ee20f0ba","Type":"ContainerStarted","Data":"cc6ddc1cfea59e40d8c120b72119e121b7779ff22cf30be7174f4f8d06cdda0d"} Jan 24 00:20:34 crc kubenswrapper[4676]: I0124 00:20:34.866753 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" event={"ID":"88763e6b-93f1-4b9d-bb0c-5c487659691a","Type":"ContainerStarted","Data":"e7b4b194ef042a28036832a666cfba7e94711668d0bfc32ab0d45edb956b6341"} Jan 24 00:20:34 crc kubenswrapper[4676]: I0124 00:20:34.885732 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvzq7" event={"ID":"3df4276b-01b4-403f-b040-85d5a8a9ef03","Type":"ContainerStarted","Data":"4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055"} Jan 24 00:20:35 crc kubenswrapper[4676]: I0124 00:20:35.895489 4676 generic.go:334] "Generic (PLEG): container finished" podID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerID="4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055" exitCode=0 Jan 24 00:20:35 crc kubenswrapper[4676]: I0124 00:20:35.895539 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvzq7" event={"ID":"3df4276b-01b4-403f-b040-85d5a8a9ef03","Type":"ContainerDied","Data":"4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055"} Jan 24 00:20:35 crc kubenswrapper[4676]: I0124 00:20:35.900060 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3bff778a-b10f-4ba9-a12f-f4086608fd30","Type":"ContainerStarted","Data":"1d968960939a9b7e1bd096a290e2ee13a9dbd2ba175fc9b933b94751653a84c4"} Jan 24 00:20:35 crc kubenswrapper[4676]: I0124 00:20:35.906343 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f3648c65-9fcb-4a9e-b4cb-d8437dc00141","Type":"ContainerStarted","Data":"8e9154fa93ad06490aa55f2202981ee1da8c0a43309a0b2dce39315135172a68"} Jan 24 00:20:35 crc kubenswrapper[4676]: I0124 00:20:35.910319 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" event={"ID":"88763e6b-93f1-4b9d-bb0c-5c487659691a","Type":"ContainerStarted","Data":"3928e7c189d053145e8953aee775f48c10df19c47e391d85edfeeb0d85f27e07"} Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.933947 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2bbaae64-ac2d-43c6-8984-5483f2eb4211","Type":"ContainerStarted","Data":"c0a4e7b27fb0d5d3a378d4a19e1e7909edf293a37101b8ea6af0ed481d778150"} Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.937470 4676 generic.go:334] "Generic (PLEG): container finished" podID="1f21d906-2974-4e56-b11f-d888e452c565" containerID="d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d" exitCode=0 Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.937572 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-sbwwk" event={"ID":"1f21d906-2974-4e56-b11f-d888e452c565","Type":"ContainerDied","Data":"d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d"} Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.942615 4676 generic.go:334] "Generic (PLEG): container finished" podID="427fbd2d-16ef-44a6-a71d-8172f56b863d" containerID="3fb2cc055bd476a6bee8209aff0183f3759ee7543512760cfe3f629dbaf87a3e" exitCode=0 Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.943309 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6sl9q" event={"ID":"427fbd2d-16ef-44a6-a71d-8172f56b863d","Type":"ContainerDied","Data":"3fb2cc055bd476a6bee8209aff0183f3759ee7543512760cfe3f629dbaf87a3e"} Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.944795 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"68d6466c-a6ff-40ba-952d-007b14efdfd3","Type":"ContainerStarted","Data":"0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723"} Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.950053 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"36558de2-6aac-43e9-832d-2f96c46e8152","Type":"ContainerStarted","Data":"9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960"} Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.956226 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c626d726-d406-41c7-8e66-07cc148d3aa7","Type":"ContainerStarted","Data":"0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62"} Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.956488 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.960734 4676 generic.go:334] "Generic (PLEG): container finished" podID="88763e6b-93f1-4b9d-bb0c-5c487659691a" containerID="3928e7c189d053145e8953aee775f48c10df19c47e391d85edfeeb0d85f27e07" exitCode=0 Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.960793 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" event={"ID":"88763e6b-93f1-4b9d-bb0c-5c487659691a","Type":"ContainerDied","Data":"3928e7c189d053145e8953aee775f48c10df19c47e391d85edfeeb0d85f27e07"} Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.966917 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jsd4n" event={"ID":"7adbcf83-efbd-4e8d-97e5-f8768463284a","Type":"ContainerStarted","Data":"aa2386ed4dd4a9df8b7b6ed99a2ba0c46a71bd82d2dedf15a8a7c0f47a77fc04"} Jan 24 00:20:36 crc kubenswrapper[4676]: I0124 00:20:36.967237 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jsd4n" Jan 24 00:20:37 crc kubenswrapper[4676]: I0124 00:20:37.018739 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=28.916833852 podStartE2EDuration="37.018718931s" podCreationTimestamp="2026-01-24 00:20:00 +0000 UTC" firstStartedPulling="2026-01-24 00:20:26.584554722 +0000 UTC m=+1010.614525713" lastFinishedPulling="2026-01-24 00:20:34.686439791 +0000 UTC m=+1018.716410792" observedRunningTime="2026-01-24 00:20:37.007615321 +0000 UTC m=+1021.037586322" watchObservedRunningTime="2026-01-24 00:20:37.018718931 +0000 UTC m=+1021.048689932" Jan 24 00:20:38 crc kubenswrapper[4676]: I0124 00:20:38.983775 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f3648c65-9fcb-4a9e-b4cb-d8437dc00141","Type":"ContainerStarted","Data":"9b2d3b0af63cc6c86788a0a84d14fea1f9e0a49f50cde256e1e22273997d0ab3"} Jan 24 00:20:38 crc kubenswrapper[4676]: I0124 00:20:38.987275 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mq6ns" event={"ID":"21550436-cd71-46d8-838e-c51c19ddf8ff","Type":"ContainerStarted","Data":"e436b0e9509a2ae3292b89ad33ee395a20f24bc93aff141bdbc6165bd9ff3e39"} Jan 24 00:20:38 crc kubenswrapper[4676]: I0124 00:20:38.990396 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-sbwwk" event={"ID":"1f21d906-2974-4e56-b11f-d888e452c565","Type":"ContainerStarted","Data":"577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be"} Jan 24 00:20:38 crc kubenswrapper[4676]: I0124 00:20:38.990845 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:38 crc kubenswrapper[4676]: I0124 00:20:38.993164 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6sl9q" event={"ID":"427fbd2d-16ef-44a6-a71d-8172f56b863d","Type":"ContainerStarted","Data":"91def523d5613035da540e2d411ee58710a4e89b8fd9300a4c87e35e1418e471"} Jan 24 00:20:38 crc kubenswrapper[4676]: I0124 00:20:38.993283 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-6sl9q" event={"ID":"427fbd2d-16ef-44a6-a71d-8172f56b863d","Type":"ContainerStarted","Data":"2c6c08b72e23b0a66e2f98c4164c911ad5d2b982503fa257694899414db35c29"} Jan 24 00:20:38 crc kubenswrapper[4676]: I0124 00:20:38.993356 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:38 crc kubenswrapper[4676]: I0124 00:20:38.993490 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.000670 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" event={"ID":"88763e6b-93f1-4b9d-bb0c-5c487659691a","Type":"ContainerStarted","Data":"d0d6ce007660b208b078333ef0f16c59301681c150c8974f5c9cb45ac7fa6ae6"} Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.000971 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.004236 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvzq7" event={"ID":"3df4276b-01b4-403f-b040-85d5a8a9ef03","Type":"ContainerStarted","Data":"23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634"} Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.011306 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"3bff778a-b10f-4ba9-a12f-f4086608fd30","Type":"ContainerStarted","Data":"99b430e598fa74475861fe6f1ff84e27aa9e920cfbaf9c8ce0d53faadf5beb79"} Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.012162 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jsd4n" podStartSLOduration=27.932326034 podStartE2EDuration="35.012145975s" podCreationTimestamp="2026-01-24 00:20:04 +0000 UTC" firstStartedPulling="2026-01-24 00:20:26.578651311 +0000 UTC m=+1010.608622312" lastFinishedPulling="2026-01-24 00:20:33.658471252 +0000 UTC m=+1017.688442253" observedRunningTime="2026-01-24 00:20:37.080494503 +0000 UTC m=+1021.110465524" watchObservedRunningTime="2026-01-24 00:20:39.012145975 +0000 UTC m=+1023.042116976" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.013495 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=19.84420677 podStartE2EDuration="33.013488545s" podCreationTimestamp="2026-01-24 00:20:06 +0000 UTC" firstStartedPulling="2026-01-24 00:20:25.219237422 +0000 UTC m=+1009.249208423" lastFinishedPulling="2026-01-24 00:20:38.388519197 +0000 UTC m=+1022.418490198" observedRunningTime="2026-01-24 00:20:39.007282885 +0000 UTC m=+1023.037253886" watchObservedRunningTime="2026-01-24 00:20:39.013488545 +0000 UTC m=+1023.043459546" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.042155 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-6sl9q" podStartSLOduration=27.155360452 podStartE2EDuration="34.042134272s" podCreationTimestamp="2026-01-24 00:20:05 +0000 UTC" firstStartedPulling="2026-01-24 00:20:26.637547144 +0000 UTC m=+1010.667518155" lastFinishedPulling="2026-01-24 00:20:33.524320954 +0000 UTC m=+1017.554291975" observedRunningTime="2026-01-24 00:20:39.034796388 +0000 UTC m=+1023.064767389" watchObservedRunningTime="2026-01-24 00:20:39.042134272 +0000 UTC m=+1023.072105273" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.062274 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-mq6ns" podStartSLOduration=5.653086369 podStartE2EDuration="10.062259089s" podCreationTimestamp="2026-01-24 00:20:29 +0000 UTC" firstStartedPulling="2026-01-24 00:20:33.957294572 +0000 UTC m=+1017.987265573" lastFinishedPulling="2026-01-24 00:20:38.366467292 +0000 UTC m=+1022.396438293" observedRunningTime="2026-01-24 00:20:39.059585217 +0000 UTC m=+1023.089556218" watchObservedRunningTime="2026-01-24 00:20:39.062259089 +0000 UTC m=+1023.092230090" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.126212 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" podStartSLOduration=9.449843346 podStartE2EDuration="10.126193177s" podCreationTimestamp="2026-01-24 00:20:29 +0000 UTC" firstStartedPulling="2026-01-24 00:20:34.051524578 +0000 UTC m=+1018.081495579" lastFinishedPulling="2026-01-24 00:20:34.727874409 +0000 UTC m=+1018.757845410" observedRunningTime="2026-01-24 00:20:39.120620616 +0000 UTC m=+1023.150591617" watchObservedRunningTime="2026-01-24 00:20:39.126193177 +0000 UTC m=+1023.156164178" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.139370 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kvzq7" podStartSLOduration=26.654079484 podStartE2EDuration="37.139353009s" podCreationTimestamp="2026-01-24 00:20:02 +0000 UTC" firstStartedPulling="2026-01-24 00:20:27.770183658 +0000 UTC m=+1011.800154659" lastFinishedPulling="2026-01-24 00:20:38.255457183 +0000 UTC m=+1022.285428184" observedRunningTime="2026-01-24 00:20:39.136072639 +0000 UTC m=+1023.166043640" watchObservedRunningTime="2026-01-24 00:20:39.139353009 +0000 UTC m=+1023.169324010" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.169861 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-sbwwk" podStartSLOduration=8.533119465 podStartE2EDuration="9.169842143s" podCreationTimestamp="2026-01-24 00:20:30 +0000 UTC" firstStartedPulling="2026-01-24 00:20:34.052059794 +0000 UTC m=+1018.082030795" lastFinishedPulling="2026-01-24 00:20:34.688782472 +0000 UTC m=+1018.718753473" observedRunningTime="2026-01-24 00:20:39.164272352 +0000 UTC m=+1023.194243353" watchObservedRunningTime="2026-01-24 00:20:39.169842143 +0000 UTC m=+1023.199813144" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.187409 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=23.390911284 podStartE2EDuration="35.18739233s" podCreationTimestamp="2026-01-24 00:20:04 +0000 UTC" firstStartedPulling="2026-01-24 00:20:26.584788149 +0000 UTC m=+1010.614759160" lastFinishedPulling="2026-01-24 00:20:38.381269195 +0000 UTC m=+1022.411240206" observedRunningTime="2026-01-24 00:20:39.182971996 +0000 UTC m=+1023.212942997" watchObservedRunningTime="2026-01-24 00:20:39.18739233 +0000 UTC m=+1023.217363331" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.223883 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:39 crc kubenswrapper[4676]: I0124 00:20:39.266927 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.021410 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.068886 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.792033 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qhgmq"] Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.842436 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-88xjq"] Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.843834 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.860174 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-88xjq"] Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.927429 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpslv\" (UniqueName: \"kubernetes.io/projected/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-kube-api-access-lpslv\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.927914 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.928010 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-config\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.928155 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:40 crc kubenswrapper[4676]: I0124 00:20:40.928247 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.035061 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.035113 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpslv\" (UniqueName: \"kubernetes.io/projected/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-kube-api-access-lpslv\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.035140 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.035171 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-config\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.035247 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.036057 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.036529 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.036598 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.037061 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-config\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.038900 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" podUID="88763e6b-93f1-4b9d-bb0c-5c487659691a" containerName="dnsmasq-dns" containerID="cri-o://d0d6ce007660b208b078333ef0f16c59301681c150c8974f5c9cb45ac7fa6ae6" gracePeriod=10 Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.052742 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.060919 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpslv\" (UniqueName: \"kubernetes.io/projected/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-kube-api-access-lpslv\") pod \"dnsmasq-dns-b8fbc5445-88xjq\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.111761 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.173314 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:41 crc kubenswrapper[4676]: I0124 00:20:41.662814 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-88xjq"] Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.020651 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.032005 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.034150 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.034837 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-6ts8l" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.034879 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.035638 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.052759 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" event={"ID":"1905ca79-a4c4-4286-8d88-2855e7b9ba4c","Type":"ContainerStarted","Data":"b390252d36e77eb3a82a5578bfec5a23efd8e93f3aaecb313b5f2fd7e776552d"} Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.055485 4676 generic.go:334] "Generic (PLEG): container finished" podID="88763e6b-93f1-4b9d-bb0c-5c487659691a" containerID="d0d6ce007660b208b078333ef0f16c59301681c150c8974f5c9cb45ac7fa6ae6" exitCode=0 Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.055767 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" event={"ID":"88763e6b-93f1-4b9d-bb0c-5c487659691a","Type":"ContainerDied","Data":"d0d6ce007660b208b078333ef0f16c59301681c150c8974f5c9cb45ac7fa6ae6"} Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.056322 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.059207 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.102672 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.153015 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4620e725-9218-461b-a56d-104bcb7f1df4-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.153054 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmrgd\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-kube-api-access-vmrgd\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.153657 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.153944 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4620e725-9218-461b-a56d-104bcb7f1df4-lock\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.154273 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4620e725-9218-461b-a56d-104bcb7f1df4-cache\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.154314 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.257488 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.257589 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4620e725-9218-461b-a56d-104bcb7f1df4-lock\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.257625 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4620e725-9218-461b-a56d-104bcb7f1df4-cache\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: E0124 00:20:42.257633 4676 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.257648 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.257674 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4620e725-9218-461b-a56d-104bcb7f1df4-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.257695 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmrgd\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-kube-api-access-vmrgd\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: E0124 00:20:42.257654 4676 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 24 00:20:42 crc kubenswrapper[4676]: E0124 00:20:42.257800 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift podName:4620e725-9218-461b-a56d-104bcb7f1df4 nodeName:}" failed. No retries permitted until 2026-01-24 00:20:42.757773062 +0000 UTC m=+1026.787744063 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift") pod "swift-storage-0" (UID: "4620e725-9218-461b-a56d-104bcb7f1df4") : configmap "swift-ring-files" not found Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.258078 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4620e725-9218-461b-a56d-104bcb7f1df4-lock\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.258086 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4620e725-9218-461b-a56d-104bcb7f1df4-cache\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.258273 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.268724 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4620e725-9218-461b-a56d-104bcb7f1df4-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.277251 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmrgd\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-kube-api-access-vmrgd\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.282834 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.567486 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.567541 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.703609 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.705102 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.709223 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Jan 24 00:20:42 crc kubenswrapper[4676]: W0124 00:20:42.709421 4676 reflector.go:561] object-"openstack"/"cert-ovnnorthd-ovndbs": failed to list *v1.Secret: secrets "cert-ovnnorthd-ovndbs" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Jan 24 00:20:42 crc kubenswrapper[4676]: E0124 00:20:42.709445 4676 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cert-ovnnorthd-ovndbs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-ovnnorthd-ovndbs\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.709985 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-g286t" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.719128 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.732530 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.766354 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1486ea92-d267-49cd-8516-d474ef25c2df-config\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.766501 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1486ea92-d267-49cd-8516-d474ef25c2df-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.766552 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.766593 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1486ea92-d267-49cd-8516-d474ef25c2df-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.766631 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2qll\" (UniqueName: \"kubernetes.io/projected/1486ea92-d267-49cd-8516-d474ef25c2df-kube-api-access-q2qll\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.766676 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1486ea92-d267-49cd-8516-d474ef25c2df-scripts\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.766701 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1486ea92-d267-49cd-8516-d474ef25c2df-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.766730 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1486ea92-d267-49cd-8516-d474ef25c2df-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: E0124 00:20:42.766973 4676 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 24 00:20:42 crc kubenswrapper[4676]: E0124 00:20:42.767506 4676 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 24 00:20:42 crc kubenswrapper[4676]: E0124 00:20:42.767762 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift podName:4620e725-9218-461b-a56d-104bcb7f1df4 nodeName:}" failed. No retries permitted until 2026-01-24 00:20:43.767743509 +0000 UTC m=+1027.797714510 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift") pod "swift-storage-0" (UID: "4620e725-9218-461b-a56d-104bcb7f1df4") : configmap "swift-ring-files" not found Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.867838 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2qll\" (UniqueName: \"kubernetes.io/projected/1486ea92-d267-49cd-8516-d474ef25c2df-kube-api-access-q2qll\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.867906 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1486ea92-d267-49cd-8516-d474ef25c2df-scripts\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.867931 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1486ea92-d267-49cd-8516-d474ef25c2df-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.867957 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1486ea92-d267-49cd-8516-d474ef25c2df-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.867975 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1486ea92-d267-49cd-8516-d474ef25c2df-config\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.868042 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1486ea92-d267-49cd-8516-d474ef25c2df-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.868077 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1486ea92-d267-49cd-8516-d474ef25c2df-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.868503 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1486ea92-d267-49cd-8516-d474ef25c2df-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.869407 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1486ea92-d267-49cd-8516-d474ef25c2df-scripts\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.870798 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1486ea92-d267-49cd-8516-d474ef25c2df-config\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.890841 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1486ea92-d267-49cd-8516-d474ef25c2df-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.891483 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1486ea92-d267-49cd-8516-d474ef25c2df-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:42 crc kubenswrapper[4676]: I0124 00:20:42.895435 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2qll\" (UniqueName: \"kubernetes.io/projected/1486ea92-d267-49cd-8516-d474ef25c2df-kube-api-access-q2qll\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:43 crc kubenswrapper[4676]: I0124 00:20:43.070442 4676 generic.go:334] "Generic (PLEG): container finished" podID="19365292-50d8-4e94-952f-2df7ee20f0ba" containerID="cc6ddc1cfea59e40d8c120b72119e121b7779ff22cf30be7174f4f8d06cdda0d" exitCode=0 Jan 24 00:20:43 crc kubenswrapper[4676]: I0124 00:20:43.070503 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"19365292-50d8-4e94-952f-2df7ee20f0ba","Type":"ContainerDied","Data":"cc6ddc1cfea59e40d8c120b72119e121b7779ff22cf30be7174f4f8d06cdda0d"} Jan 24 00:20:43 crc kubenswrapper[4676]: I0124 00:20:43.075629 4676 generic.go:334] "Generic (PLEG): container finished" podID="2bbaae64-ac2d-43c6-8984-5483f2eb4211" containerID="c0a4e7b27fb0d5d3a378d4a19e1e7909edf293a37101b8ea6af0ed481d778150" exitCode=0 Jan 24 00:20:43 crc kubenswrapper[4676]: I0124 00:20:43.076105 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2bbaae64-ac2d-43c6-8984-5483f2eb4211","Type":"ContainerDied","Data":"c0a4e7b27fb0d5d3a378d4a19e1e7909edf293a37101b8ea6af0ed481d778150"} Jan 24 00:20:43 crc kubenswrapper[4676]: I0124 00:20:43.622575 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kvzq7" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerName="registry-server" probeResult="failure" output=< Jan 24 00:20:43 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:20:43 crc kubenswrapper[4676]: > Jan 24 00:20:43 crc kubenswrapper[4676]: I0124 00:20:43.779474 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:43 crc kubenswrapper[4676]: E0124 00:20:43.779639 4676 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 24 00:20:43 crc kubenswrapper[4676]: E0124 00:20:43.779665 4676 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 24 00:20:43 crc kubenswrapper[4676]: E0124 00:20:43.779732 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift podName:4620e725-9218-461b-a56d-104bcb7f1df4 nodeName:}" failed. No retries permitted until 2026-01-24 00:20:45.779709778 +0000 UTC m=+1029.809680789 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift") pod "swift-storage-0" (UID: "4620e725-9218-461b-a56d-104bcb7f1df4") : configmap "swift-ring-files" not found Jan 24 00:20:43 crc kubenswrapper[4676]: I0124 00:20:43.815573 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Jan 24 00:20:43 crc kubenswrapper[4676]: I0124 00:20:43.825935 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/1486ea92-d267-49cd-8516-d474ef25c2df-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"1486ea92-d267-49cd-8516-d474ef25c2df\") " pod="openstack/ovn-northd-0" Jan 24 00:20:43 crc kubenswrapper[4676]: I0124 00:20:43.960488 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.105411 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"19365292-50d8-4e94-952f-2df7ee20f0ba","Type":"ContainerStarted","Data":"e6c89b124d66353c91ee1aa9ed1ccee9c8d066f5478a92d37d4d5862729ffec2"} Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.126415 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2bbaae64-ac2d-43c6-8984-5483f2eb4211","Type":"ContainerStarted","Data":"25d79390127c82f1424411b19f4ba1ae769f32f2036b81b6574e43dee5a1980a"} Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.135466 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=41.086598672 podStartE2EDuration="47.135450092s" podCreationTimestamp="2026-01-24 00:19:57 +0000 UTC" firstStartedPulling="2026-01-24 00:20:26.637701799 +0000 UTC m=+1010.667672800" lastFinishedPulling="2026-01-24 00:20:32.686553229 +0000 UTC m=+1016.716524220" observedRunningTime="2026-01-24 00:20:44.13085249 +0000 UTC m=+1028.160823491" watchObservedRunningTime="2026-01-24 00:20:44.135450092 +0000 UTC m=+1028.165421093" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.162297 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371988.692497 podStartE2EDuration="48.162279163s" podCreationTimestamp="2026-01-24 00:19:56 +0000 UTC" firstStartedPulling="2026-01-24 00:20:10.695446279 +0000 UTC m=+994.725417310" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:20:44.157763265 +0000 UTC m=+1028.187734266" watchObservedRunningTime="2026-01-24 00:20:44.162279163 +0000 UTC m=+1028.192250164" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.474849 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.477896 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:44 crc kubenswrapper[4676]: W0124 00:20:44.482625 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1486ea92_d267_49cd_8516_d474ef25c2df.slice/crio-184c22321dc7f5e257d71d3ac4bd871cbe4616dccc9cc9b42dad26475bb88d2e WatchSource:0}: Error finding container 184c22321dc7f5e257d71d3ac4bd871cbe4616dccc9cc9b42dad26475bb88d2e: Status 404 returned error can't find the container with id 184c22321dc7f5e257d71d3ac4bd871cbe4616dccc9cc9b42dad26475bb88d2e Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.594566 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-config\") pod \"88763e6b-93f1-4b9d-bb0c-5c487659691a\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.594610 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx7ts\" (UniqueName: \"kubernetes.io/projected/88763e6b-93f1-4b9d-bb0c-5c487659691a-kube-api-access-gx7ts\") pod \"88763e6b-93f1-4b9d-bb0c-5c487659691a\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.594700 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-ovsdbserver-nb\") pod \"88763e6b-93f1-4b9d-bb0c-5c487659691a\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.594798 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-dns-svc\") pod \"88763e6b-93f1-4b9d-bb0c-5c487659691a\" (UID: \"88763e6b-93f1-4b9d-bb0c-5c487659691a\") " Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.629541 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88763e6b-93f1-4b9d-bb0c-5c487659691a-kube-api-access-gx7ts" (OuterVolumeSpecName: "kube-api-access-gx7ts") pod "88763e6b-93f1-4b9d-bb0c-5c487659691a" (UID: "88763e6b-93f1-4b9d-bb0c-5c487659691a"). InnerVolumeSpecName "kube-api-access-gx7ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.656527 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-config" (OuterVolumeSpecName: "config") pod "88763e6b-93f1-4b9d-bb0c-5c487659691a" (UID: "88763e6b-93f1-4b9d-bb0c-5c487659691a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.664633 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "88763e6b-93f1-4b9d-bb0c-5c487659691a" (UID: "88763e6b-93f1-4b9d-bb0c-5c487659691a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.674866 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "88763e6b-93f1-4b9d-bb0c-5c487659691a" (UID: "88763e6b-93f1-4b9d-bb0c-5c487659691a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.696182 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.696215 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.696228 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx7ts\" (UniqueName: \"kubernetes.io/projected/88763e6b-93f1-4b9d-bb0c-5c487659691a-kube-api-access-gx7ts\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:44 crc kubenswrapper[4676]: I0124 00:20:44.696241 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88763e6b-93f1-4b9d-bb0c-5c487659691a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.134727 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" event={"ID":"88763e6b-93f1-4b9d-bb0c-5c487659691a","Type":"ContainerDied","Data":"e7b4b194ef042a28036832a666cfba7e94711668d0bfc32ab0d45edb956b6341"} Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.135010 4676 scope.go:117] "RemoveContainer" containerID="d0d6ce007660b208b078333ef0f16c59301681c150c8974f5c9cb45ac7fa6ae6" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.134781 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-qhgmq" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.136780 4676 generic.go:334] "Generic (PLEG): container finished" podID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" containerID="837bffae0bd65cfec1fbd0e46c064b367c74a4b7d5a87b7e24ab722d34c34bd4" exitCode=0 Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.136864 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" event={"ID":"1905ca79-a4c4-4286-8d88-2855e7b9ba4c","Type":"ContainerDied","Data":"837bffae0bd65cfec1fbd0e46c064b367c74a4b7d5a87b7e24ab722d34c34bd4"} Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.139880 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1486ea92-d267-49cd-8516-d474ef25c2df","Type":"ContainerStarted","Data":"184c22321dc7f5e257d71d3ac4bd871cbe4616dccc9cc9b42dad26475bb88d2e"} Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.152974 4676 scope.go:117] "RemoveContainer" containerID="3928e7c189d053145e8953aee775f48c10df19c47e391d85edfeeb0d85f27e07" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.307584 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qhgmq"] Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.313483 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-qhgmq"] Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.714591 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.821348 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:45 crc kubenswrapper[4676]: E0124 00:20:45.823065 4676 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 24 00:20:45 crc kubenswrapper[4676]: E0124 00:20:45.823093 4676 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 24 00:20:45 crc kubenswrapper[4676]: E0124 00:20:45.823137 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift podName:4620e725-9218-461b-a56d-104bcb7f1df4 nodeName:}" failed. No retries permitted until 2026-01-24 00:20:49.823122142 +0000 UTC m=+1033.853093143 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift") pod "swift-storage-0" (UID: "4620e725-9218-461b-a56d-104bcb7f1df4") : configmap "swift-ring-files" not found Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.885892 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-5fmzb"] Jan 24 00:20:45 crc kubenswrapper[4676]: E0124 00:20:45.886198 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88763e6b-93f1-4b9d-bb0c-5c487659691a" containerName="init" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.886209 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="88763e6b-93f1-4b9d-bb0c-5c487659691a" containerName="init" Jan 24 00:20:45 crc kubenswrapper[4676]: E0124 00:20:45.886237 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88763e6b-93f1-4b9d-bb0c-5c487659691a" containerName="dnsmasq-dns" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.886243 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="88763e6b-93f1-4b9d-bb0c-5c487659691a" containerName="dnsmasq-dns" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.886394 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="88763e6b-93f1-4b9d-bb0c-5c487659691a" containerName="dnsmasq-dns" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.886824 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.898611 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.902720 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.905811 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Jan 24 00:20:45 crc kubenswrapper[4676]: I0124 00:20:45.911728 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-5fmzb"] Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.024333 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-combined-ca-bundle\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.024393 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zd5k\" (UniqueName: \"kubernetes.io/projected/4ff61f48-e451-47e8-adcc-0870b29d28a9-kube-api-access-2zd5k\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.024434 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ff61f48-e451-47e8-adcc-0870b29d28a9-etc-swift\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.024459 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-ring-data-devices\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.024502 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-scripts\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.024531 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-swiftconf\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.024563 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-dispersionconf\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.126107 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zd5k\" (UniqueName: \"kubernetes.io/projected/4ff61f48-e451-47e8-adcc-0870b29d28a9-kube-api-access-2zd5k\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.126406 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ff61f48-e451-47e8-adcc-0870b29d28a9-etc-swift\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.126434 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-ring-data-devices\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.126482 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-scripts\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.126514 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-swiftconf\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.126545 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-dispersionconf\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.126601 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-combined-ca-bundle\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.127056 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-ring-data-devices\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.127300 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ff61f48-e451-47e8-adcc-0870b29d28a9-etc-swift\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.127775 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-scripts\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.130809 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-combined-ca-bundle\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.131518 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-dispersionconf\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.131594 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-swiftconf\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.146018 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zd5k\" (UniqueName: \"kubernetes.io/projected/4ff61f48-e451-47e8-adcc-0870b29d28a9-kube-api-access-2zd5k\") pod \"swift-ring-rebalance-5fmzb\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.147035 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" event={"ID":"1905ca79-a4c4-4286-8d88-2855e7b9ba4c","Type":"ContainerStarted","Data":"1d6d0279f417026b6ba6bd78b2b2fd59bc265f683401ca01652ddd44d6b071a2"} Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.147238 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.168622 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" podStartSLOduration=6.168606171 podStartE2EDuration="6.168606171s" podCreationTimestamp="2026-01-24 00:20:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:20:46.163167375 +0000 UTC m=+1030.193138376" watchObservedRunningTime="2026-01-24 00:20:46.168606171 +0000 UTC m=+1030.198577172" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.219405 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.278965 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88763e6b-93f1-4b9d-bb0c-5c487659691a" path="/var/lib/kubelet/pods/88763e6b-93f1-4b9d-bb0c-5c487659691a/volumes" Jan 24 00:20:46 crc kubenswrapper[4676]: W0124 00:20:46.720593 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ff61f48_e451_47e8_adcc_0870b29d28a9.slice/crio-edae868ed3f37d7cd803c42347470b3ff0f9857923570cef0ccdd12a7ade323b WatchSource:0}: Error finding container edae868ed3f37d7cd803c42347470b3ff0f9857923570cef0ccdd12a7ade323b: Status 404 returned error can't find the container with id edae868ed3f37d7cd803c42347470b3ff0f9857923570cef0ccdd12a7ade323b Jan 24 00:20:46 crc kubenswrapper[4676]: I0124 00:20:46.721534 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-5fmzb"] Jan 24 00:20:47 crc kubenswrapper[4676]: I0124 00:20:47.162101 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1486ea92-d267-49cd-8516-d474ef25c2df","Type":"ContainerStarted","Data":"2bc81f0c9e99c0a686f722f292e67c5c773c0f09646441418ba3c8b48bf7b43d"} Jan 24 00:20:47 crc kubenswrapper[4676]: I0124 00:20:47.162148 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"1486ea92-d267-49cd-8516-d474ef25c2df","Type":"ContainerStarted","Data":"0d6819a80061f8595b56a96c4f3467bde91ad0c094f9c254dd9411a388f0a563"} Jan 24 00:20:47 crc kubenswrapper[4676]: I0124 00:20:47.162202 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Jan 24 00:20:47 crc kubenswrapper[4676]: I0124 00:20:47.163049 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5fmzb" event={"ID":"4ff61f48-e451-47e8-adcc-0870b29d28a9","Type":"ContainerStarted","Data":"edae868ed3f37d7cd803c42347470b3ff0f9857923570cef0ccdd12a7ade323b"} Jan 24 00:20:47 crc kubenswrapper[4676]: I0124 00:20:47.182815 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.717620582 podStartE2EDuration="5.182799748s" podCreationTimestamp="2026-01-24 00:20:42 +0000 UTC" firstStartedPulling="2026-01-24 00:20:44.485292465 +0000 UTC m=+1028.515263466" lastFinishedPulling="2026-01-24 00:20:45.950471631 +0000 UTC m=+1029.980442632" observedRunningTime="2026-01-24 00:20:47.178661022 +0000 UTC m=+1031.208632083" watchObservedRunningTime="2026-01-24 00:20:47.182799748 +0000 UTC m=+1031.212770749" Jan 24 00:20:47 crc kubenswrapper[4676]: I0124 00:20:47.432019 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Jan 24 00:20:47 crc kubenswrapper[4676]: I0124 00:20:47.432088 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Jan 24 00:20:48 crc kubenswrapper[4676]: I0124 00:20:48.639335 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Jan 24 00:20:48 crc kubenswrapper[4676]: I0124 00:20:48.639706 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Jan 24 00:20:49 crc kubenswrapper[4676]: I0124 00:20:49.747489 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Jan 24 00:20:49 crc kubenswrapper[4676]: I0124 00:20:49.828834 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Jan 24 00:20:49 crc kubenswrapper[4676]: I0124 00:20:49.900211 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:49 crc kubenswrapper[4676]: E0124 00:20:49.901127 4676 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 24 00:20:49 crc kubenswrapper[4676]: E0124 00:20:49.901147 4676 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 24 00:20:49 crc kubenswrapper[4676]: E0124 00:20:49.901184 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift podName:4620e725-9218-461b-a56d-104bcb7f1df4 nodeName:}" failed. No retries permitted until 2026-01-24 00:20:57.901169691 +0000 UTC m=+1041.931140692 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift") pod "swift-storage-0" (UID: "4620e725-9218-461b-a56d-104bcb7f1df4") : configmap "swift-ring-files" not found Jan 24 00:20:50 crc kubenswrapper[4676]: I0124 00:20:50.676626 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 24 00:20:50 crc kubenswrapper[4676]: I0124 00:20:50.953905 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.042789 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.174519 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.193183 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5fmzb" event={"ID":"4ff61f48-e451-47e8-adcc-0870b29d28a9","Type":"ContainerStarted","Data":"a2edcf50a838124e8087fc219bf5e4c2e5444fe834157a78116049634218ec2a"} Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.228839 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-5fmzb" podStartSLOduration=2.221291726 podStartE2EDuration="6.228822497s" podCreationTimestamp="2026-01-24 00:20:45 +0000 UTC" firstStartedPulling="2026-01-24 00:20:46.722804612 +0000 UTC m=+1030.752775623" lastFinishedPulling="2026-01-24 00:20:50.730335392 +0000 UTC m=+1034.760306394" observedRunningTime="2026-01-24 00:20:51.218615955 +0000 UTC m=+1035.248586956" watchObservedRunningTime="2026-01-24 00:20:51.228822497 +0000 UTC m=+1035.258793498" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.247013 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-sbwwk"] Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.247450 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-sbwwk" podUID="1f21d906-2974-4e56-b11f-d888e452c565" containerName="dnsmasq-dns" containerID="cri-o://577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be" gracePeriod=10 Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.666769 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.739187 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btl5m\" (UniqueName: \"kubernetes.io/projected/1f21d906-2974-4e56-b11f-d888e452c565-kube-api-access-btl5m\") pod \"1f21d906-2974-4e56-b11f-d888e452c565\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.739297 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-nb\") pod \"1f21d906-2974-4e56-b11f-d888e452c565\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.739334 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-sb\") pod \"1f21d906-2974-4e56-b11f-d888e452c565\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.740100 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-dns-svc\") pod \"1f21d906-2974-4e56-b11f-d888e452c565\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.740117 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-config\") pod \"1f21d906-2974-4e56-b11f-d888e452c565\" (UID: \"1f21d906-2974-4e56-b11f-d888e452c565\") " Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.745273 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f21d906-2974-4e56-b11f-d888e452c565-kube-api-access-btl5m" (OuterVolumeSpecName: "kube-api-access-btl5m") pod "1f21d906-2974-4e56-b11f-d888e452c565" (UID: "1f21d906-2974-4e56-b11f-d888e452c565"). InnerVolumeSpecName "kube-api-access-btl5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.792451 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-config" (OuterVolumeSpecName: "config") pod "1f21d906-2974-4e56-b11f-d888e452c565" (UID: "1f21d906-2974-4e56-b11f-d888e452c565"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.794928 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1f21d906-2974-4e56-b11f-d888e452c565" (UID: "1f21d906-2974-4e56-b11f-d888e452c565"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.810221 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1f21d906-2974-4e56-b11f-d888e452c565" (UID: "1f21d906-2974-4e56-b11f-d888e452c565"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.843444 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.843469 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.843478 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btl5m\" (UniqueName: \"kubernetes.io/projected/1f21d906-2974-4e56-b11f-d888e452c565-kube-api-access-btl5m\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.843511 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.852785 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1f21d906-2974-4e56-b11f-d888e452c565" (UID: "1f21d906-2974-4e56-b11f-d888e452c565"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:51 crc kubenswrapper[4676]: I0124 00:20:51.945175 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f21d906-2974-4e56-b11f-d888e452c565-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.199764 4676 generic.go:334] "Generic (PLEG): container finished" podID="1f21d906-2974-4e56-b11f-d888e452c565" containerID="577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be" exitCode=0 Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.200465 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-sbwwk" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.200564 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-sbwwk" event={"ID":"1f21d906-2974-4e56-b11f-d888e452c565","Type":"ContainerDied","Data":"577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be"} Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.200586 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-sbwwk" event={"ID":"1f21d906-2974-4e56-b11f-d888e452c565","Type":"ContainerDied","Data":"089f8e4a35ef84364abbbae1879307f5831ba1a07f81a4231c6cfc6207ff2e6a"} Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.200601 4676 scope.go:117] "RemoveContainer" containerID="577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.229348 4676 scope.go:117] "RemoveContainer" containerID="d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.254358 4676 scope.go:117] "RemoveContainer" containerID="577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be" Jan 24 00:20:52 crc kubenswrapper[4676]: E0124 00:20:52.255786 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be\": container with ID starting with 577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be not found: ID does not exist" containerID="577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.255814 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be"} err="failed to get container status \"577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be\": rpc error: code = NotFound desc = could not find container \"577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be\": container with ID starting with 577f50aa5b4437d91afcb97c82c41e9f5ee408e1ba9e1be7852f62684447d5be not found: ID does not exist" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.255839 4676 scope.go:117] "RemoveContainer" containerID="d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d" Jan 24 00:20:52 crc kubenswrapper[4676]: E0124 00:20:52.256039 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d\": container with ID starting with d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d not found: ID does not exist" containerID="d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.256057 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d"} err="failed to get container status \"d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d\": rpc error: code = NotFound desc = could not find container \"d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d\": container with ID starting with d33ce991197b97fff8c8aedf2ecfa7d98c5e7a4c61013dd5a3b38abbab942c0d not found: ID does not exist" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.267235 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-sbwwk"] Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.267277 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-sbwwk"] Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.609035 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.650928 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:52 crc kubenswrapper[4676]: I0124 00:20:52.842664 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kvzq7"] Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.077237 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-d7f6-account-create-update-brcr8"] Jan 24 00:20:54 crc kubenswrapper[4676]: E0124 00:20:54.077803 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f21d906-2974-4e56-b11f-d888e452c565" containerName="dnsmasq-dns" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.077814 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f21d906-2974-4e56-b11f-d888e452c565" containerName="dnsmasq-dns" Jan 24 00:20:54 crc kubenswrapper[4676]: E0124 00:20:54.077831 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f21d906-2974-4e56-b11f-d888e452c565" containerName="init" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.077837 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f21d906-2974-4e56-b11f-d888e452c565" containerName="init" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.077997 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f21d906-2974-4e56-b11f-d888e452c565" containerName="dnsmasq-dns" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.078494 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.080759 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.105956 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d7f6-account-create-update-brcr8"] Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.122636 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-ph986"] Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.123874 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ph986" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.153744 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ph986"] Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.181216 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-operator-scripts\") pod \"glance-d7f6-account-create-update-brcr8\" (UID: \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\") " pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.181474 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssl97\" (UniqueName: \"kubernetes.io/projected/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-kube-api-access-ssl97\") pod \"glance-db-create-ph986\" (UID: \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\") " pod="openstack/glance-db-create-ph986" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.181598 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-operator-scripts\") pod \"glance-db-create-ph986\" (UID: \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\") " pod="openstack/glance-db-create-ph986" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.181670 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md5qp\" (UniqueName: \"kubernetes.io/projected/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-kube-api-access-md5qp\") pod \"glance-d7f6-account-create-update-brcr8\" (UID: \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\") " pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.213327 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kvzq7" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerName="registry-server" containerID="cri-o://23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634" gracePeriod=2 Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.268086 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f21d906-2974-4e56-b11f-d888e452c565" path="/var/lib/kubelet/pods/1f21d906-2974-4e56-b11f-d888e452c565/volumes" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.284500 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssl97\" (UniqueName: \"kubernetes.io/projected/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-kube-api-access-ssl97\") pod \"glance-db-create-ph986\" (UID: \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\") " pod="openstack/glance-db-create-ph986" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.285550 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-operator-scripts\") pod \"glance-db-create-ph986\" (UID: \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\") " pod="openstack/glance-db-create-ph986" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.285681 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md5qp\" (UniqueName: \"kubernetes.io/projected/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-kube-api-access-md5qp\") pod \"glance-d7f6-account-create-update-brcr8\" (UID: \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\") " pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.285782 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-operator-scripts\") pod \"glance-d7f6-account-create-update-brcr8\" (UID: \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\") " pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.286978 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-operator-scripts\") pod \"glance-d7f6-account-create-update-brcr8\" (UID: \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\") " pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.287111 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-operator-scripts\") pod \"glance-db-create-ph986\" (UID: \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\") " pod="openstack/glance-db-create-ph986" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.317475 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md5qp\" (UniqueName: \"kubernetes.io/projected/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-kube-api-access-md5qp\") pod \"glance-d7f6-account-create-update-brcr8\" (UID: \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\") " pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.320460 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssl97\" (UniqueName: \"kubernetes.io/projected/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-kube-api-access-ssl97\") pod \"glance-db-create-ph986\" (UID: \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\") " pod="openstack/glance-db-create-ph986" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.400296 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.448757 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ph986" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.615839 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.695425 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-utilities\") pod \"3df4276b-01b4-403f-b040-85d5a8a9ef03\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.695587 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-catalog-content\") pod \"3df4276b-01b4-403f-b040-85d5a8a9ef03\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.695614 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk82l\" (UniqueName: \"kubernetes.io/projected/3df4276b-01b4-403f-b040-85d5a8a9ef03-kube-api-access-tk82l\") pod \"3df4276b-01b4-403f-b040-85d5a8a9ef03\" (UID: \"3df4276b-01b4-403f-b040-85d5a8a9ef03\") " Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.697659 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-utilities" (OuterVolumeSpecName: "utilities") pod "3df4276b-01b4-403f-b040-85d5a8a9ef03" (UID: "3df4276b-01b4-403f-b040-85d5a8a9ef03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.699561 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.702197 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3df4276b-01b4-403f-b040-85d5a8a9ef03-kube-api-access-tk82l" (OuterVolumeSpecName: "kube-api-access-tk82l") pod "3df4276b-01b4-403f-b040-85d5a8a9ef03" (UID: "3df4276b-01b4-403f-b040-85d5a8a9ef03"). InnerVolumeSpecName "kube-api-access-tk82l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.759706 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3df4276b-01b4-403f-b040-85d5a8a9ef03" (UID: "3df4276b-01b4-403f-b040-85d5a8a9ef03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.800837 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df4276b-01b4-403f-b040-85d5a8a9ef03-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.801087 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk82l\" (UniqueName: \"kubernetes.io/projected/3df4276b-01b4-403f-b040-85d5a8a9ef03-kube-api-access-tk82l\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:54 crc kubenswrapper[4676]: I0124 00:20:54.892704 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-d7f6-account-create-update-brcr8"] Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.024040 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-ph986"] Jan 24 00:20:55 crc kubenswrapper[4676]: W0124 00:20:55.026073 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod925957f4_9ca1_4e46_b3ac_0dd0ca5713cd.slice/crio-00a687de5e1dc9cd4413233520312cb969bf584cd8c200af2ba5dc7b9782e4f4 WatchSource:0}: Error finding container 00a687de5e1dc9cd4413233520312cb969bf584cd8c200af2ba5dc7b9782e4f4: Status 404 returned error can't find the container with id 00a687de5e1dc9cd4413233520312cb969bf584cd8c200af2ba5dc7b9782e4f4 Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.223651 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ph986" event={"ID":"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd","Type":"ContainerStarted","Data":"a55ff63829260fcafaf5bd727871713740d9ed36c94c4fc218cf31c8bb4fa53b"} Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.223731 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ph986" event={"ID":"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd","Type":"ContainerStarted","Data":"00a687de5e1dc9cd4413233520312cb969bf584cd8c200af2ba5dc7b9782e4f4"} Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.227398 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d7f6-account-create-update-brcr8" event={"ID":"98c25d86-7dc6-4b15-9511-3a7cbf4c7592","Type":"ContainerStarted","Data":"8280acdb35983168359f77fef7c062161338aa47a5bf092b620c75a0dd83736f"} Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.227436 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d7f6-account-create-update-brcr8" event={"ID":"98c25d86-7dc6-4b15-9511-3a7cbf4c7592","Type":"ContainerStarted","Data":"520e95ec58419aeecc535479889b038fb2987a491e8f1735dc7b52a21a89b798"} Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.232237 4676 generic.go:334] "Generic (PLEG): container finished" podID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerID="23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634" exitCode=0 Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.232286 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvzq7" event={"ID":"3df4276b-01b4-403f-b040-85d5a8a9ef03","Type":"ContainerDied","Data":"23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634"} Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.232527 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kvzq7" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.232914 4676 scope.go:117] "RemoveContainer" containerID="23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.232309 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kvzq7" event={"ID":"3df4276b-01b4-403f-b040-85d5a8a9ef03","Type":"ContainerDied","Data":"88c849f82d3193bf091f531d74d4fde6ab5977e0c495ee6e86efeb15c6348494"} Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.256655 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-ph986" podStartSLOduration=1.256636728 podStartE2EDuration="1.256636728s" podCreationTimestamp="2026-01-24 00:20:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:20:55.242498375 +0000 UTC m=+1039.272469376" watchObservedRunningTime="2026-01-24 00:20:55.256636728 +0000 UTC m=+1039.286607729" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.263562 4676 scope.go:117] "RemoveContainer" containerID="4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.272881 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-d7f6-account-create-update-brcr8" podStartSLOduration=1.272864005 podStartE2EDuration="1.272864005s" podCreationTimestamp="2026-01-24 00:20:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:20:55.268892184 +0000 UTC m=+1039.298863175" watchObservedRunningTime="2026-01-24 00:20:55.272864005 +0000 UTC m=+1039.302835006" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.289425 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kvzq7"] Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.294735 4676 scope.go:117] "RemoveContainer" containerID="82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.317734 4676 scope.go:117] "RemoveContainer" containerID="23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.318615 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kvzq7"] Jan 24 00:20:55 crc kubenswrapper[4676]: E0124 00:20:55.320971 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634\": container with ID starting with 23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634 not found: ID does not exist" containerID="23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.321003 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634"} err="failed to get container status \"23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634\": rpc error: code = NotFound desc = could not find container \"23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634\": container with ID starting with 23f51327ad99c429533212b03ced853abc5e0f0d71914be27d57c58e73611634 not found: ID does not exist" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.321025 4676 scope.go:117] "RemoveContainer" containerID="4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055" Jan 24 00:20:55 crc kubenswrapper[4676]: E0124 00:20:55.321341 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055\": container with ID starting with 4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055 not found: ID does not exist" containerID="4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.321361 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055"} err="failed to get container status \"4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055\": rpc error: code = NotFound desc = could not find container \"4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055\": container with ID starting with 4f4e99bafbb1a85396c9179c36382959eaa8fbb41e0c1ee24af103d0f0f87055 not found: ID does not exist" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.321400 4676 scope.go:117] "RemoveContainer" containerID="82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e" Jan 24 00:20:55 crc kubenswrapper[4676]: E0124 00:20:55.321710 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e\": container with ID starting with 82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e not found: ID does not exist" containerID="82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e" Jan 24 00:20:55 crc kubenswrapper[4676]: I0124 00:20:55.321742 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e"} err="failed to get container status \"82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e\": rpc error: code = NotFound desc = could not find container \"82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e\": container with ID starting with 82d05dea3899aabc4e3fbab8a0c2cac7fba0b90a8cfd043aada7232a1d18050e not found: ID does not exist" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.101994 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-5zwrz"] Jan 24 00:20:56 crc kubenswrapper[4676]: E0124 00:20:56.102315 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerName="registry-server" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.102335 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerName="registry-server" Jan 24 00:20:56 crc kubenswrapper[4676]: E0124 00:20:56.102461 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerName="extract-utilities" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.102474 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerName="extract-utilities" Jan 24 00:20:56 crc kubenswrapper[4676]: E0124 00:20:56.102488 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerName="extract-content" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.102502 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerName="extract-content" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.102671 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" containerName="registry-server" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.103223 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5zwrz" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.107551 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.121773 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5zwrz"] Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.265753 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/320759e3-3504-4bd4-b55a-b8cb664e1ce0-operator-scripts\") pod \"root-account-create-update-5zwrz\" (UID: \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\") " pod="openstack/root-account-create-update-5zwrz" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.265811 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krzzm\" (UniqueName: \"kubernetes.io/projected/320759e3-3504-4bd4-b55a-b8cb664e1ce0-kube-api-access-krzzm\") pod \"root-account-create-update-5zwrz\" (UID: \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\") " pod="openstack/root-account-create-update-5zwrz" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.287252 4676 generic.go:334] "Generic (PLEG): container finished" podID="925957f4-9ca1-4e46-b3ac-0dd0ca5713cd" containerID="a55ff63829260fcafaf5bd727871713740d9ed36c94c4fc218cf31c8bb4fa53b" exitCode=0 Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.288518 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3df4276b-01b4-403f-b040-85d5a8a9ef03" path="/var/lib/kubelet/pods/3df4276b-01b4-403f-b040-85d5a8a9ef03/volumes" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.289112 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ph986" event={"ID":"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd","Type":"ContainerDied","Data":"a55ff63829260fcafaf5bd727871713740d9ed36c94c4fc218cf31c8bb4fa53b"} Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.292865 4676 generic.go:334] "Generic (PLEG): container finished" podID="98c25d86-7dc6-4b15-9511-3a7cbf4c7592" containerID="8280acdb35983168359f77fef7c062161338aa47a5bf092b620c75a0dd83736f" exitCode=0 Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.293001 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d7f6-account-create-update-brcr8" event={"ID":"98c25d86-7dc6-4b15-9511-3a7cbf4c7592","Type":"ContainerDied","Data":"8280acdb35983168359f77fef7c062161338aa47a5bf092b620c75a0dd83736f"} Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.366827 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/320759e3-3504-4bd4-b55a-b8cb664e1ce0-operator-scripts\") pod \"root-account-create-update-5zwrz\" (UID: \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\") " pod="openstack/root-account-create-update-5zwrz" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.366870 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krzzm\" (UniqueName: \"kubernetes.io/projected/320759e3-3504-4bd4-b55a-b8cb664e1ce0-kube-api-access-krzzm\") pod \"root-account-create-update-5zwrz\" (UID: \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\") " pod="openstack/root-account-create-update-5zwrz" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.368070 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/320759e3-3504-4bd4-b55a-b8cb664e1ce0-operator-scripts\") pod \"root-account-create-update-5zwrz\" (UID: \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\") " pod="openstack/root-account-create-update-5zwrz" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.390203 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krzzm\" (UniqueName: \"kubernetes.io/projected/320759e3-3504-4bd4-b55a-b8cb664e1ce0-kube-api-access-krzzm\") pod \"root-account-create-update-5zwrz\" (UID: \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\") " pod="openstack/root-account-create-update-5zwrz" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.421976 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5zwrz" Jan 24 00:20:56 crc kubenswrapper[4676]: I0124 00:20:56.859446 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-5zwrz"] Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.308254 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5zwrz" event={"ID":"320759e3-3504-4bd4-b55a-b8cb664e1ce0","Type":"ContainerStarted","Data":"d2ed0ffdbd6bb9c785556f969ad9aeabbea68ca6c96b8f79735c762793724bfe"} Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.308305 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5zwrz" event={"ID":"320759e3-3504-4bd4-b55a-b8cb664e1ce0","Type":"ContainerStarted","Data":"a1c7b5b20a313f8f253066a409ba7019da788bdf5e0a133113d60074a31eb7cf"} Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.332914 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-5zwrz" podStartSLOduration=1.332893149 podStartE2EDuration="1.332893149s" podCreationTimestamp="2026-01-24 00:20:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:20:57.324594215 +0000 UTC m=+1041.354565226" watchObservedRunningTime="2026-01-24 00:20:57.332893149 +0000 UTC m=+1041.362864150" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.703124 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ph986" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.707977 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.790006 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssl97\" (UniqueName: \"kubernetes.io/projected/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-kube-api-access-ssl97\") pod \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\" (UID: \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\") " Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.790094 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-operator-scripts\") pod \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\" (UID: \"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd\") " Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.790137 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-operator-scripts\") pod \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\" (UID: \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\") " Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.790268 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-md5qp\" (UniqueName: \"kubernetes.io/projected/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-kube-api-access-md5qp\") pod \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\" (UID: \"98c25d86-7dc6-4b15-9511-3a7cbf4c7592\") " Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.790670 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "925957f4-9ca1-4e46-b3ac-0dd0ca5713cd" (UID: "925957f4-9ca1-4e46-b3ac-0dd0ca5713cd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.790746 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "98c25d86-7dc6-4b15-9511-3a7cbf4c7592" (UID: "98c25d86-7dc6-4b15-9511-3a7cbf4c7592"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.791549 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.791575 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.795864 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-kube-api-access-md5qp" (OuterVolumeSpecName: "kube-api-access-md5qp") pod "98c25d86-7dc6-4b15-9511-3a7cbf4c7592" (UID: "98c25d86-7dc6-4b15-9511-3a7cbf4c7592"). InnerVolumeSpecName "kube-api-access-md5qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.796084 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-kube-api-access-ssl97" (OuterVolumeSpecName: "kube-api-access-ssl97") pod "925957f4-9ca1-4e46-b3ac-0dd0ca5713cd" (UID: "925957f4-9ca1-4e46-b3ac-0dd0ca5713cd"). InnerVolumeSpecName "kube-api-access-ssl97". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.892494 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-md5qp\" (UniqueName: \"kubernetes.io/projected/98c25d86-7dc6-4b15-9511-3a7cbf4c7592-kube-api-access-md5qp\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.892519 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssl97\" (UniqueName: \"kubernetes.io/projected/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd-kube-api-access-ssl97\") on node \"crc\" DevicePath \"\"" Jan 24 00:20:57 crc kubenswrapper[4676]: I0124 00:20:57.994538 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:20:57 crc kubenswrapper[4676]: E0124 00:20:57.994736 4676 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 24 00:20:57 crc kubenswrapper[4676]: E0124 00:20:57.994762 4676 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 24 00:20:57 crc kubenswrapper[4676]: E0124 00:20:57.994820 4676 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift podName:4620e725-9218-461b-a56d-104bcb7f1df4 nodeName:}" failed. No retries permitted until 2026-01-24 00:21:13.994803028 +0000 UTC m=+1058.024774029 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift") pod "swift-storage-0" (UID: "4620e725-9218-461b-a56d-104bcb7f1df4") : configmap "swift-ring-files" not found Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.336243 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-ph986" event={"ID":"925957f4-9ca1-4e46-b3ac-0dd0ca5713cd","Type":"ContainerDied","Data":"00a687de5e1dc9cd4413233520312cb969bf584cd8c200af2ba5dc7b9782e4f4"} Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.336288 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00a687de5e1dc9cd4413233520312cb969bf584cd8c200af2ba5dc7b9782e4f4" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.336369 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-ph986" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.340956 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-d7f6-account-create-update-brcr8" event={"ID":"98c25d86-7dc6-4b15-9511-3a7cbf4c7592","Type":"ContainerDied","Data":"520e95ec58419aeecc535479889b038fb2987a491e8f1735dc7b52a21a89b798"} Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.340996 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="520e95ec58419aeecc535479889b038fb2987a491e8f1735dc7b52a21a89b798" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.341097 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-d7f6-account-create-update-brcr8" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.347354 4676 generic.go:334] "Generic (PLEG): container finished" podID="320759e3-3504-4bd4-b55a-b8cb664e1ce0" containerID="d2ed0ffdbd6bb9c785556f969ad9aeabbea68ca6c96b8f79735c762793724bfe" exitCode=0 Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.347465 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5zwrz" event={"ID":"320759e3-3504-4bd4-b55a-b8cb664e1ce0","Type":"ContainerDied","Data":"d2ed0ffdbd6bb9c785556f969ad9aeabbea68ca6c96b8f79735c762793724bfe"} Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.422290 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-x4ptm"] Jan 24 00:20:58 crc kubenswrapper[4676]: E0124 00:20:58.422607 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c25d86-7dc6-4b15-9511-3a7cbf4c7592" containerName="mariadb-account-create-update" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.422619 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c25d86-7dc6-4b15-9511-3a7cbf4c7592" containerName="mariadb-account-create-update" Jan 24 00:20:58 crc kubenswrapper[4676]: E0124 00:20:58.422647 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="925957f4-9ca1-4e46-b3ac-0dd0ca5713cd" containerName="mariadb-database-create" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.422654 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="925957f4-9ca1-4e46-b3ac-0dd0ca5713cd" containerName="mariadb-database-create" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.422792 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c25d86-7dc6-4b15-9511-3a7cbf4c7592" containerName="mariadb-account-create-update" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.422810 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="925957f4-9ca1-4e46-b3ac-0dd0ca5713cd" containerName="mariadb-database-create" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.423289 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x4ptm" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.435912 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-x4ptm"] Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.507166 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgccq\" (UniqueName: \"kubernetes.io/projected/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-kube-api-access-zgccq\") pod \"keystone-db-create-x4ptm\" (UID: \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\") " pod="openstack/keystone-db-create-x4ptm" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.507221 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-operator-scripts\") pod \"keystone-db-create-x4ptm\" (UID: \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\") " pod="openstack/keystone-db-create-x4ptm" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.534831 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-60c3-account-create-update-xz4kb"] Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.535703 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.537433 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.542243 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-60c3-account-create-update-xz4kb"] Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.609309 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-operator-scripts\") pod \"keystone-db-create-x4ptm\" (UID: \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\") " pod="openstack/keystone-db-create-x4ptm" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.609631 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-operator-scripts\") pod \"keystone-60c3-account-create-update-xz4kb\" (UID: \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\") " pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.609650 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2xkg\" (UniqueName: \"kubernetes.io/projected/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-kube-api-access-z2xkg\") pod \"keystone-60c3-account-create-update-xz4kb\" (UID: \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\") " pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.609717 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgccq\" (UniqueName: \"kubernetes.io/projected/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-kube-api-access-zgccq\") pod \"keystone-db-create-x4ptm\" (UID: \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\") " pod="openstack/keystone-db-create-x4ptm" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.610428 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-operator-scripts\") pod \"keystone-db-create-x4ptm\" (UID: \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\") " pod="openstack/keystone-db-create-x4ptm" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.633323 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgccq\" (UniqueName: \"kubernetes.io/projected/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-kube-api-access-zgccq\") pod \"keystone-db-create-x4ptm\" (UID: \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\") " pod="openstack/keystone-db-create-x4ptm" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.711132 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-operator-scripts\") pod \"keystone-60c3-account-create-update-xz4kb\" (UID: \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\") " pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.711395 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2xkg\" (UniqueName: \"kubernetes.io/projected/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-kube-api-access-z2xkg\") pod \"keystone-60c3-account-create-update-xz4kb\" (UID: \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\") " pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.711817 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-operator-scripts\") pod \"keystone-60c3-account-create-update-xz4kb\" (UID: \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\") " pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.736954 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2xkg\" (UniqueName: \"kubernetes.io/projected/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-kube-api-access-z2xkg\") pod \"keystone-60c3-account-create-update-xz4kb\" (UID: \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\") " pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.737959 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x4ptm" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.782949 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-prlkx"] Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.783877 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-prlkx" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.793597 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-prlkx"] Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.851430 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.920174 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-operator-scripts\") pod \"placement-db-create-prlkx\" (UID: \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\") " pod="openstack/placement-db-create-prlkx" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.920513 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf6r7\" (UniqueName: \"kubernetes.io/projected/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-kube-api-access-tf6r7\") pod \"placement-db-create-prlkx\" (UID: \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\") " pod="openstack/placement-db-create-prlkx" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.921464 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-278e-account-create-update-8fwkd"] Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.922538 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.929756 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Jan 24 00:20:58 crc kubenswrapper[4676]: I0124 00:20:58.941915 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-278e-account-create-update-8fwkd"] Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.022760 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf6r7\" (UniqueName: \"kubernetes.io/projected/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-kube-api-access-tf6r7\") pod \"placement-db-create-prlkx\" (UID: \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\") " pod="openstack/placement-db-create-prlkx" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.022828 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439dce27-2a59-49a2-beab-82482c1ce3cb-operator-scripts\") pod \"placement-278e-account-create-update-8fwkd\" (UID: \"439dce27-2a59-49a2-beab-82482c1ce3cb\") " pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.022871 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-operator-scripts\") pod \"placement-db-create-prlkx\" (UID: \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\") " pod="openstack/placement-db-create-prlkx" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.022923 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bt4p\" (UniqueName: \"kubernetes.io/projected/439dce27-2a59-49a2-beab-82482c1ce3cb-kube-api-access-9bt4p\") pod \"placement-278e-account-create-update-8fwkd\" (UID: \"439dce27-2a59-49a2-beab-82482c1ce3cb\") " pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.025971 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-operator-scripts\") pod \"placement-db-create-prlkx\" (UID: \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\") " pod="openstack/placement-db-create-prlkx" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.046089 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf6r7\" (UniqueName: \"kubernetes.io/projected/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-kube-api-access-tf6r7\") pod \"placement-db-create-prlkx\" (UID: \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\") " pod="openstack/placement-db-create-prlkx" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.049011 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.125463 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bt4p\" (UniqueName: \"kubernetes.io/projected/439dce27-2a59-49a2-beab-82482c1ce3cb-kube-api-access-9bt4p\") pod \"placement-278e-account-create-update-8fwkd\" (UID: \"439dce27-2a59-49a2-beab-82482c1ce3cb\") " pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.125642 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439dce27-2a59-49a2-beab-82482c1ce3cb-operator-scripts\") pod \"placement-278e-account-create-update-8fwkd\" (UID: \"439dce27-2a59-49a2-beab-82482c1ce3cb\") " pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.126478 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439dce27-2a59-49a2-beab-82482c1ce3cb-operator-scripts\") pod \"placement-278e-account-create-update-8fwkd\" (UID: \"439dce27-2a59-49a2-beab-82482c1ce3cb\") " pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.147838 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bt4p\" (UniqueName: \"kubernetes.io/projected/439dce27-2a59-49a2-beab-82482c1ce3cb-kube-api-access-9bt4p\") pod \"placement-278e-account-create-update-8fwkd\" (UID: \"439dce27-2a59-49a2-beab-82482c1ce3cb\") " pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.148805 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-prlkx" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.275829 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.278125 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-x4ptm"] Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.356682 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-kbtsc"] Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.357652 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.360340 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.360557 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-s5zsb" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.377915 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-x4ptm" event={"ID":"d6d8987f-6f37-4773-99ad-6bc7fd8971f2","Type":"ContainerStarted","Data":"5e11825261a30b8eb14d5f7d2fc6206facb16e2daf33a5873c566c87690bb5b5"} Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.379238 4676 generic.go:334] "Generic (PLEG): container finished" podID="4ff61f48-e451-47e8-adcc-0870b29d28a9" containerID="a2edcf50a838124e8087fc219bf5e4c2e5444fe834157a78116049634218ec2a" exitCode=0 Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.379396 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5fmzb" event={"ID":"4ff61f48-e451-47e8-adcc-0870b29d28a9","Type":"ContainerDied","Data":"a2edcf50a838124e8087fc219bf5e4c2e5444fe834157a78116049634218ec2a"} Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.379477 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kbtsc"] Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.432046 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-combined-ca-bundle\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.432133 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6qmf\" (UniqueName: \"kubernetes.io/projected/7ebdc261-7c0d-49a2-9233-868fda906788-kube-api-access-t6qmf\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.432177 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-db-sync-config-data\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.432215 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-config-data\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.442490 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-60c3-account-create-update-xz4kb"] Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.538146 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6qmf\" (UniqueName: \"kubernetes.io/projected/7ebdc261-7c0d-49a2-9233-868fda906788-kube-api-access-t6qmf\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.538220 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-db-sync-config-data\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.538263 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-config-data\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.538321 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-combined-ca-bundle\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.544680 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-db-sync-config-data\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.545690 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-config-data\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.546938 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-combined-ca-bundle\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.567068 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6qmf\" (UniqueName: \"kubernetes.io/projected/7ebdc261-7c0d-49a2-9233-868fda906788-kube-api-access-t6qmf\") pod \"glance-db-sync-kbtsc\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.691038 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-prlkx"] Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.747550 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kbtsc" Jan 24 00:20:59 crc kubenswrapper[4676]: I0124 00:20:59.845303 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-278e-account-create-update-8fwkd"] Jan 24 00:20:59 crc kubenswrapper[4676]: W0124 00:20:59.909405 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod439dce27_2a59_49a2_beab_82482c1ce3cb.slice/crio-db36ca4b04238f99f0a64b033baa2442b0b23d7197b6ba5c71dd6b5dc07b0500 WatchSource:0}: Error finding container db36ca4b04238f99f0a64b033baa2442b0b23d7197b6ba5c71dd6b5dc07b0500: Status 404 returned error can't find the container with id db36ca4b04238f99f0a64b033baa2442b0b23d7197b6ba5c71dd6b5dc07b0500 Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.092930 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5zwrz" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.152909 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/320759e3-3504-4bd4-b55a-b8cb664e1ce0-operator-scripts\") pod \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\" (UID: \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\") " Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.153059 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krzzm\" (UniqueName: \"kubernetes.io/projected/320759e3-3504-4bd4-b55a-b8cb664e1ce0-kube-api-access-krzzm\") pod \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\" (UID: \"320759e3-3504-4bd4-b55a-b8cb664e1ce0\") " Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.155189 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/320759e3-3504-4bd4-b55a-b8cb664e1ce0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "320759e3-3504-4bd4-b55a-b8cb664e1ce0" (UID: "320759e3-3504-4bd4-b55a-b8cb664e1ce0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.195610 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/320759e3-3504-4bd4-b55a-b8cb664e1ce0-kube-api-access-krzzm" (OuterVolumeSpecName: "kube-api-access-krzzm") pod "320759e3-3504-4bd4-b55a-b8cb664e1ce0" (UID: "320759e3-3504-4bd4-b55a-b8cb664e1ce0"). InnerVolumeSpecName "kube-api-access-krzzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.258466 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krzzm\" (UniqueName: \"kubernetes.io/projected/320759e3-3504-4bd4-b55a-b8cb664e1ce0-kube-api-access-krzzm\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.258494 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/320759e3-3504-4bd4-b55a-b8cb664e1ce0-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.310817 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-kbtsc"] Jan 24 00:21:00 crc kubenswrapper[4676]: W0124 00:21:00.321675 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ebdc261_7c0d_49a2_9233_868fda906788.slice/crio-e532ae412cda92f465aa9e050762bb063a13c20d49e7ab63fe45cd1ee4ddc01b WatchSource:0}: Error finding container e532ae412cda92f465aa9e050762bb063a13c20d49e7ab63fe45cd1ee4ddc01b: Status 404 returned error can't find the container with id e532ae412cda92f465aa9e050762bb063a13c20d49e7ab63fe45cd1ee4ddc01b Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.392387 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kbtsc" event={"ID":"7ebdc261-7c0d-49a2-9233-868fda906788","Type":"ContainerStarted","Data":"e532ae412cda92f465aa9e050762bb063a13c20d49e7ab63fe45cd1ee4ddc01b"} Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.393734 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-278e-account-create-update-8fwkd" event={"ID":"439dce27-2a59-49a2-beab-82482c1ce3cb","Type":"ContainerStarted","Data":"b1e8f656f78cf6d79b93f73d693ca8f011b12dbd49906ddf8afe723c5036938e"} Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.393772 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-278e-account-create-update-8fwkd" event={"ID":"439dce27-2a59-49a2-beab-82482c1ce3cb","Type":"ContainerStarted","Data":"db36ca4b04238f99f0a64b033baa2442b0b23d7197b6ba5c71dd6b5dc07b0500"} Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.395288 4676 generic.go:334] "Generic (PLEG): container finished" podID="4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42" containerID="7b1df20731b13a3b81d6ecd68e39c0140b372b18913afe443cbe4c6b00f28ea4" exitCode=0 Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.395445 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-60c3-account-create-update-xz4kb" event={"ID":"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42","Type":"ContainerDied","Data":"7b1df20731b13a3b81d6ecd68e39c0140b372b18913afe443cbe4c6b00f28ea4"} Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.395478 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-60c3-account-create-update-xz4kb" event={"ID":"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42","Type":"ContainerStarted","Data":"10f1971dd827090a805df3cf92e6f04a60f7b6bf074806ac40846fa7212ad001"} Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.398319 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-5zwrz" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.398322 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-5zwrz" event={"ID":"320759e3-3504-4bd4-b55a-b8cb664e1ce0","Type":"ContainerDied","Data":"a1c7b5b20a313f8f253066a409ba7019da788bdf5e0a133113d60074a31eb7cf"} Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.398582 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1c7b5b20a313f8f253066a409ba7019da788bdf5e0a133113d60074a31eb7cf" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.399753 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-prlkx" event={"ID":"2374740d-d0b0-4a0c-b23e-d41d3ad524ac","Type":"ContainerStarted","Data":"d5c280bb67bfb03306230881277a4f3cf73d70180a6643cdcdc6c29c26f5cdcb"} Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.399779 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-prlkx" event={"ID":"2374740d-d0b0-4a0c-b23e-d41d3ad524ac","Type":"ContainerStarted","Data":"2431e513e27795629ea214c200cf83d2b16cbdbd751217fe08aac5b0be029497"} Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.401581 4676 generic.go:334] "Generic (PLEG): container finished" podID="d6d8987f-6f37-4773-99ad-6bc7fd8971f2" containerID="3555dab39dcc24cb3d738eb810520147594f4722ea764e376bb893954691a29d" exitCode=0 Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.401650 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-x4ptm" event={"ID":"d6d8987f-6f37-4773-99ad-6bc7fd8971f2","Type":"ContainerDied","Data":"3555dab39dcc24cb3d738eb810520147594f4722ea764e376bb893954691a29d"} Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.415635 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-278e-account-create-update-8fwkd" podStartSLOduration=2.415616728 podStartE2EDuration="2.415616728s" podCreationTimestamp="2026-01-24 00:20:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:00.410609635 +0000 UTC m=+1044.440580636" watchObservedRunningTime="2026-01-24 00:21:00.415616728 +0000 UTC m=+1044.445587729" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.776360 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.797512 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-prlkx" podStartSLOduration=2.797497223 podStartE2EDuration="2.797497223s" podCreationTimestamp="2026-01-24 00:20:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:00.490614646 +0000 UTC m=+1044.520585647" watchObservedRunningTime="2026-01-24 00:21:00.797497223 +0000 UTC m=+1044.827468214" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.869267 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-ring-data-devices\") pod \"4ff61f48-e451-47e8-adcc-0870b29d28a9\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.869312 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zd5k\" (UniqueName: \"kubernetes.io/projected/4ff61f48-e451-47e8-adcc-0870b29d28a9-kube-api-access-2zd5k\") pod \"4ff61f48-e451-47e8-adcc-0870b29d28a9\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.869345 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-swiftconf\") pod \"4ff61f48-e451-47e8-adcc-0870b29d28a9\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.869407 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-scripts\") pod \"4ff61f48-e451-47e8-adcc-0870b29d28a9\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.869468 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-dispersionconf\") pod \"4ff61f48-e451-47e8-adcc-0870b29d28a9\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.869492 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-combined-ca-bundle\") pod \"4ff61f48-e451-47e8-adcc-0870b29d28a9\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.869652 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ff61f48-e451-47e8-adcc-0870b29d28a9-etc-swift\") pod \"4ff61f48-e451-47e8-adcc-0870b29d28a9\" (UID: \"4ff61f48-e451-47e8-adcc-0870b29d28a9\") " Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.870967 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "4ff61f48-e451-47e8-adcc-0870b29d28a9" (UID: "4ff61f48-e451-47e8-adcc-0870b29d28a9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.871359 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ff61f48-e451-47e8-adcc-0870b29d28a9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4ff61f48-e451-47e8-adcc-0870b29d28a9" (UID: "4ff61f48-e451-47e8-adcc-0870b29d28a9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.874284 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ff61f48-e451-47e8-adcc-0870b29d28a9-kube-api-access-2zd5k" (OuterVolumeSpecName: "kube-api-access-2zd5k") pod "4ff61f48-e451-47e8-adcc-0870b29d28a9" (UID: "4ff61f48-e451-47e8-adcc-0870b29d28a9"). InnerVolumeSpecName "kube-api-access-2zd5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.881369 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "4ff61f48-e451-47e8-adcc-0870b29d28a9" (UID: "4ff61f48-e451-47e8-adcc-0870b29d28a9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.898537 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-scripts" (OuterVolumeSpecName: "scripts") pod "4ff61f48-e451-47e8-adcc-0870b29d28a9" (UID: "4ff61f48-e451-47e8-adcc-0870b29d28a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.898757 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ff61f48-e451-47e8-adcc-0870b29d28a9" (UID: "4ff61f48-e451-47e8-adcc-0870b29d28a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.900185 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "4ff61f48-e451-47e8-adcc-0870b29d28a9" (UID: "4ff61f48-e451-47e8-adcc-0870b29d28a9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.971616 4676 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/4ff61f48-e451-47e8-adcc-0870b29d28a9-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.971648 4676 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.971660 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zd5k\" (UniqueName: \"kubernetes.io/projected/4ff61f48-e451-47e8-adcc-0870b29d28a9-kube-api-access-2zd5k\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.971669 4676 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.971677 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff61f48-e451-47e8-adcc-0870b29d28a9-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.971685 4676 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:00 crc kubenswrapper[4676]: I0124 00:21:00.971694 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ff61f48-e451-47e8-adcc-0870b29d28a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.410915 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-5fmzb" Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.410956 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-5fmzb" event={"ID":"4ff61f48-e451-47e8-adcc-0870b29d28a9","Type":"ContainerDied","Data":"edae868ed3f37d7cd803c42347470b3ff0f9857923570cef0ccdd12a7ade323b"} Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.410992 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edae868ed3f37d7cd803c42347470b3ff0f9857923570cef0ccdd12a7ade323b" Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.412568 4676 generic.go:334] "Generic (PLEG): container finished" podID="439dce27-2a59-49a2-beab-82482c1ce3cb" containerID="b1e8f656f78cf6d79b93f73d693ca8f011b12dbd49906ddf8afe723c5036938e" exitCode=0 Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.412631 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-278e-account-create-update-8fwkd" event={"ID":"439dce27-2a59-49a2-beab-82482c1ce3cb","Type":"ContainerDied","Data":"b1e8f656f78cf6d79b93f73d693ca8f011b12dbd49906ddf8afe723c5036938e"} Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.415048 4676 generic.go:334] "Generic (PLEG): container finished" podID="2374740d-d0b0-4a0c-b23e-d41d3ad524ac" containerID="d5c280bb67bfb03306230881277a4f3cf73d70180a6643cdcdc6c29c26f5cdcb" exitCode=0 Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.415284 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-prlkx" event={"ID":"2374740d-d0b0-4a0c-b23e-d41d3ad524ac","Type":"ContainerDied","Data":"d5c280bb67bfb03306230881277a4f3cf73d70180a6643cdcdc6c29c26f5cdcb"} Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.891291 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.895757 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x4ptm" Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.988982 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgccq\" (UniqueName: \"kubernetes.io/projected/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-kube-api-access-zgccq\") pod \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\" (UID: \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\") " Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.989028 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-operator-scripts\") pod \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\" (UID: \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\") " Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.989087 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-operator-scripts\") pod \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\" (UID: \"d6d8987f-6f37-4773-99ad-6bc7fd8971f2\") " Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.989166 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2xkg\" (UniqueName: \"kubernetes.io/projected/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-kube-api-access-z2xkg\") pod \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\" (UID: \"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42\") " Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.990601 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d6d8987f-6f37-4773-99ad-6bc7fd8971f2" (UID: "d6d8987f-6f37-4773-99ad-6bc7fd8971f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.992598 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42" (UID: "4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.995168 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-kube-api-access-zgccq" (OuterVolumeSpecName: "kube-api-access-zgccq") pod "d6d8987f-6f37-4773-99ad-6bc7fd8971f2" (UID: "d6d8987f-6f37-4773-99ad-6bc7fd8971f2"). InnerVolumeSpecName "kube-api-access-zgccq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:01 crc kubenswrapper[4676]: I0124 00:21:01.996072 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-kube-api-access-z2xkg" (OuterVolumeSpecName: "kube-api-access-z2xkg") pod "4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42" (UID: "4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42"). InnerVolumeSpecName "kube-api-access-z2xkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.091021 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgccq\" (UniqueName: \"kubernetes.io/projected/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-kube-api-access-zgccq\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.091053 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.091067 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d6d8987f-6f37-4773-99ad-6bc7fd8971f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.091078 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2xkg\" (UniqueName: \"kubernetes.io/projected/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42-kube-api-access-z2xkg\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.245323 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-5zwrz"] Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.252118 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-5zwrz"] Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.273446 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="320759e3-3504-4bd4-b55a-b8cb664e1ce0" path="/var/lib/kubelet/pods/320759e3-3504-4bd4-b55a-b8cb664e1ce0/volumes" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.425049 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-x4ptm" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.425081 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-x4ptm" event={"ID":"d6d8987f-6f37-4773-99ad-6bc7fd8971f2","Type":"ContainerDied","Data":"5e11825261a30b8eb14d5f7d2fc6206facb16e2daf33a5873c566c87690bb5b5"} Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.425118 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e11825261a30b8eb14d5f7d2fc6206facb16e2daf33a5873c566c87690bb5b5" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.430320 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-60c3-account-create-update-xz4kb" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.430411 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-60c3-account-create-update-xz4kb" event={"ID":"4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42","Type":"ContainerDied","Data":"10f1971dd827090a805df3cf92e6f04a60f7b6bf074806ac40846fa7212ad001"} Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.430457 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10f1971dd827090a805df3cf92e6f04a60f7b6bf074806ac40846fa7212ad001" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.766606 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-prlkx" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.903244 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf6r7\" (UniqueName: \"kubernetes.io/projected/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-kube-api-access-tf6r7\") pod \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\" (UID: \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\") " Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.903343 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-operator-scripts\") pod \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\" (UID: \"2374740d-d0b0-4a0c-b23e-d41d3ad524ac\") " Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.904094 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2374740d-d0b0-4a0c-b23e-d41d3ad524ac" (UID: "2374740d-d0b0-4a0c-b23e-d41d3ad524ac"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.928020 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-kube-api-access-tf6r7" (OuterVolumeSpecName: "kube-api-access-tf6r7") pod "2374740d-d0b0-4a0c-b23e-d41d3ad524ac" (UID: "2374740d-d0b0-4a0c-b23e-d41d3ad524ac"). InnerVolumeSpecName "kube-api-access-tf6r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:02 crc kubenswrapper[4676]: I0124 00:21:02.966726 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.004890 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf6r7\" (UniqueName: \"kubernetes.io/projected/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-kube-api-access-tf6r7\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.004919 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2374740d-d0b0-4a0c-b23e-d41d3ad524ac-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.105752 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bt4p\" (UniqueName: \"kubernetes.io/projected/439dce27-2a59-49a2-beab-82482c1ce3cb-kube-api-access-9bt4p\") pod \"439dce27-2a59-49a2-beab-82482c1ce3cb\" (UID: \"439dce27-2a59-49a2-beab-82482c1ce3cb\") " Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.105881 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439dce27-2a59-49a2-beab-82482c1ce3cb-operator-scripts\") pod \"439dce27-2a59-49a2-beab-82482c1ce3cb\" (UID: \"439dce27-2a59-49a2-beab-82482c1ce3cb\") " Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.106692 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/439dce27-2a59-49a2-beab-82482c1ce3cb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "439dce27-2a59-49a2-beab-82482c1ce3cb" (UID: "439dce27-2a59-49a2-beab-82482c1ce3cb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.106887 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/439dce27-2a59-49a2-beab-82482c1ce3cb-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.130797 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/439dce27-2a59-49a2-beab-82482c1ce3cb-kube-api-access-9bt4p" (OuterVolumeSpecName: "kube-api-access-9bt4p") pod "439dce27-2a59-49a2-beab-82482c1ce3cb" (UID: "439dce27-2a59-49a2-beab-82482c1ce3cb"). InnerVolumeSpecName "kube-api-access-9bt4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.208960 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bt4p\" (UniqueName: \"kubernetes.io/projected/439dce27-2a59-49a2-beab-82482c1ce3cb-kube-api-access-9bt4p\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.439264 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-prlkx" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.439260 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-prlkx" event={"ID":"2374740d-d0b0-4a0c-b23e-d41d3ad524ac","Type":"ContainerDied","Data":"2431e513e27795629ea214c200cf83d2b16cbdbd751217fe08aac5b0be029497"} Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.439604 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2431e513e27795629ea214c200cf83d2b16cbdbd751217fe08aac5b0be029497" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.440615 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-278e-account-create-update-8fwkd" event={"ID":"439dce27-2a59-49a2-beab-82482c1ce3cb","Type":"ContainerDied","Data":"db36ca4b04238f99f0a64b033baa2442b0b23d7197b6ba5c71dd6b5dc07b0500"} Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.440636 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db36ca4b04238f99f0a64b033baa2442b0b23d7197b6ba5c71dd6b5dc07b0500" Jan 24 00:21:03 crc kubenswrapper[4676]: I0124 00:21:03.440683 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-278e-account-create-update-8fwkd" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.256801 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-drjg7"] Jan 24 00:21:07 crc kubenswrapper[4676]: E0124 00:21:07.257323 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="320759e3-3504-4bd4-b55a-b8cb664e1ce0" containerName="mariadb-account-create-update" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257334 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="320759e3-3504-4bd4-b55a-b8cb664e1ce0" containerName="mariadb-account-create-update" Jan 24 00:21:07 crc kubenswrapper[4676]: E0124 00:21:07.257349 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="439dce27-2a59-49a2-beab-82482c1ce3cb" containerName="mariadb-account-create-update" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257355 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="439dce27-2a59-49a2-beab-82482c1ce3cb" containerName="mariadb-account-create-update" Jan 24 00:21:07 crc kubenswrapper[4676]: E0124 00:21:07.257368 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff61f48-e451-47e8-adcc-0870b29d28a9" containerName="swift-ring-rebalance" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257388 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff61f48-e451-47e8-adcc-0870b29d28a9" containerName="swift-ring-rebalance" Jan 24 00:21:07 crc kubenswrapper[4676]: E0124 00:21:07.257401 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2374740d-d0b0-4a0c-b23e-d41d3ad524ac" containerName="mariadb-database-create" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257407 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2374740d-d0b0-4a0c-b23e-d41d3ad524ac" containerName="mariadb-database-create" Jan 24 00:21:07 crc kubenswrapper[4676]: E0124 00:21:07.257426 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42" containerName="mariadb-account-create-update" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257433 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42" containerName="mariadb-account-create-update" Jan 24 00:21:07 crc kubenswrapper[4676]: E0124 00:21:07.257443 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6d8987f-6f37-4773-99ad-6bc7fd8971f2" containerName="mariadb-database-create" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257449 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6d8987f-6f37-4773-99ad-6bc7fd8971f2" containerName="mariadb-database-create" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257580 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42" containerName="mariadb-account-create-update" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257592 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="320759e3-3504-4bd4-b55a-b8cb664e1ce0" containerName="mariadb-account-create-update" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257601 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ff61f48-e451-47e8-adcc-0870b29d28a9" containerName="swift-ring-rebalance" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257609 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6d8987f-6f37-4773-99ad-6bc7fd8971f2" containerName="mariadb-database-create" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257619 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="439dce27-2a59-49a2-beab-82482c1ce3cb" containerName="mariadb-account-create-update" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.257625 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2374740d-d0b0-4a0c-b23e-d41d3ad524ac" containerName="mariadb-database-create" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.258148 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.260776 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.264534 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-drjg7"] Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.391298 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20421570-a66e-4903-937f-af2ab847b5f1-operator-scripts\") pod \"root-account-create-update-drjg7\" (UID: \"20421570-a66e-4903-937f-af2ab847b5f1\") " pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.392213 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7wkn\" (UniqueName: \"kubernetes.io/projected/20421570-a66e-4903-937f-af2ab847b5f1-kube-api-access-b7wkn\") pod \"root-account-create-update-drjg7\" (UID: \"20421570-a66e-4903-937f-af2ab847b5f1\") " pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.493496 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20421570-a66e-4903-937f-af2ab847b5f1-operator-scripts\") pod \"root-account-create-update-drjg7\" (UID: \"20421570-a66e-4903-937f-af2ab847b5f1\") " pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.493545 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7wkn\" (UniqueName: \"kubernetes.io/projected/20421570-a66e-4903-937f-af2ab847b5f1-kube-api-access-b7wkn\") pod \"root-account-create-update-drjg7\" (UID: \"20421570-a66e-4903-937f-af2ab847b5f1\") " pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.494149 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20421570-a66e-4903-937f-af2ab847b5f1-operator-scripts\") pod \"root-account-create-update-drjg7\" (UID: \"20421570-a66e-4903-937f-af2ab847b5f1\") " pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.528226 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7wkn\" (UniqueName: \"kubernetes.io/projected/20421570-a66e-4903-937f-af2ab847b5f1-kube-api-access-b7wkn\") pod \"root-account-create-update-drjg7\" (UID: \"20421570-a66e-4903-937f-af2ab847b5f1\") " pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:07 crc kubenswrapper[4676]: I0124 00:21:07.575224 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:08 crc kubenswrapper[4676]: I0124 00:21:08.480299 4676 generic.go:334] "Generic (PLEG): container finished" podID="36558de2-6aac-43e9-832d-2f96c46e8152" containerID="9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960" exitCode=0 Jan 24 00:21:08 crc kubenswrapper[4676]: I0124 00:21:08.480439 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"36558de2-6aac-43e9-832d-2f96c46e8152","Type":"ContainerDied","Data":"9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960"} Jan 24 00:21:08 crc kubenswrapper[4676]: I0124 00:21:08.482455 4676 generic.go:334] "Generic (PLEG): container finished" podID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerID="0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723" exitCode=0 Jan 24 00:21:08 crc kubenswrapper[4676]: I0124 00:21:08.482502 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"68d6466c-a6ff-40ba-952d-007b14efdfd3","Type":"ContainerDied","Data":"0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723"} Jan 24 00:21:10 crc kubenswrapper[4676]: I0124 00:21:10.380490 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jsd4n" podUID="7adbcf83-efbd-4e8d-97e5-f8768463284a" containerName="ovn-controller" probeResult="failure" output=< Jan 24 00:21:10 crc kubenswrapper[4676]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 24 00:21:10 crc kubenswrapper[4676]: > Jan 24 00:21:10 crc kubenswrapper[4676]: I0124 00:21:10.503784 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:21:10 crc kubenswrapper[4676]: I0124 00:21:10.504653 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-6sl9q" Jan 24 00:21:10 crc kubenswrapper[4676]: I0124 00:21:10.942447 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jsd4n-config-lcgr8"] Jan 24 00:21:10 crc kubenswrapper[4676]: I0124 00:21:10.945521 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:10 crc kubenswrapper[4676]: I0124 00:21:10.947717 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 24 00:21:10 crc kubenswrapper[4676]: I0124 00:21:10.952041 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jsd4n-config-lcgr8"] Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.054903 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run-ovn\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.055046 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-log-ovn\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.055077 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8s74\" (UniqueName: \"kubernetes.io/projected/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-kube-api-access-w8s74\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.055535 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-scripts\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.055766 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-additional-scripts\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.055796 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.157764 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-additional-scripts\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.158120 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.158155 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run-ovn\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.158243 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-log-ovn\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.158271 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8s74\" (UniqueName: \"kubernetes.io/projected/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-kube-api-access-w8s74\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.158337 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-scripts\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.158390 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.158445 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-additional-scripts\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.158454 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-log-ovn\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.158492 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run-ovn\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.161021 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-scripts\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.196460 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8s74\" (UniqueName: \"kubernetes.io/projected/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-kube-api-access-w8s74\") pod \"ovn-controller-jsd4n-config-lcgr8\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:11 crc kubenswrapper[4676]: I0124 00:21:11.267771 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:14 crc kubenswrapper[4676]: I0124 00:21:14.004642 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:21:14 crc kubenswrapper[4676]: I0124 00:21:14.018098 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4620e725-9218-461b-a56d-104bcb7f1df4-etc-swift\") pod \"swift-storage-0\" (UID: \"4620e725-9218-461b-a56d-104bcb7f1df4\") " pod="openstack/swift-storage-0" Jan 24 00:21:14 crc kubenswrapper[4676]: I0124 00:21:14.151544 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 24 00:21:15 crc kubenswrapper[4676]: I0124 00:21:15.387192 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jsd4n" podUID="7adbcf83-efbd-4e8d-97e5-f8768463284a" containerName="ovn-controller" probeResult="failure" output=< Jan 24 00:21:15 crc kubenswrapper[4676]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 24 00:21:15 crc kubenswrapper[4676]: > Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.134417 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-drjg7"] Jan 24 00:21:16 crc kubenswrapper[4676]: W0124 00:21:16.146461 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20421570_a66e_4903_937f_af2ab847b5f1.slice/crio-6fe7e9ee134a3384cdb14a9c06f284d65f4ca15260ccd9bddf8bd062136f714f WatchSource:0}: Error finding container 6fe7e9ee134a3384cdb14a9c06f284d65f4ca15260ccd9bddf8bd062136f714f: Status 404 returned error can't find the container with id 6fe7e9ee134a3384cdb14a9c06f284d65f4ca15260ccd9bddf8bd062136f714f Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.242750 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jsd4n-config-lcgr8"] Jan 24 00:21:16 crc kubenswrapper[4676]: W0124 00:21:16.254847 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5533b3f_acbd_48c8_ad6f_96ba8ea7a435.slice/crio-6ab4ddbfca70e6160be9f44ee5e62ac3fe6c06d281ba30970058467edbdbafdb WatchSource:0}: Error finding container 6ab4ddbfca70e6160be9f44ee5e62ac3fe6c06d281ba30970058467edbdbafdb: Status 404 returned error can't find the container with id 6ab4ddbfca70e6160be9f44ee5e62ac3fe6c06d281ba30970058467edbdbafdb Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.360978 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.555235 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"68d6466c-a6ff-40ba-952d-007b14efdfd3","Type":"ContainerStarted","Data":"7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385"} Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.555594 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.556985 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"36558de2-6aac-43e9-832d-2f96c46e8152","Type":"ContainerStarted","Data":"028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9"} Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.557579 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.558823 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kbtsc" event={"ID":"7ebdc261-7c0d-49a2-9233-868fda906788","Type":"ContainerStarted","Data":"1c08827f5b622dc0d2e57d1709fbff0d00bb4f69592994ac4ce101a373fb04cb"} Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.559915 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jsd4n-config-lcgr8" event={"ID":"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435","Type":"ContainerStarted","Data":"6ab4ddbfca70e6160be9f44ee5e62ac3fe6c06d281ba30970058467edbdbafdb"} Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.560702 4676 generic.go:334] "Generic (PLEG): container finished" podID="20421570-a66e-4903-937f-af2ab847b5f1" containerID="a94f61fa91f9d6c0d5e0b0e07bb624e5cb428e8002a06af7cee89d3dcb90c4cd" exitCode=0 Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.560739 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drjg7" event={"ID":"20421570-a66e-4903-937f-af2ab847b5f1","Type":"ContainerDied","Data":"a94f61fa91f9d6c0d5e0b0e07bb624e5cb428e8002a06af7cee89d3dcb90c4cd"} Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.560753 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drjg7" event={"ID":"20421570-a66e-4903-937f-af2ab847b5f1","Type":"ContainerStarted","Data":"6fe7e9ee134a3384cdb14a9c06f284d65f4ca15260ccd9bddf8bd062136f714f"} Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.561371 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"76813373ab91f4561b4c34edc63c14ee6c3985d8e49f0b8588af9a13cafe7bed"} Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.589936 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=45.023265926 podStartE2EDuration="1m22.589921505s" podCreationTimestamp="2026-01-24 00:19:54 +0000 UTC" firstStartedPulling="2026-01-24 00:19:56.385922389 +0000 UTC m=+980.415893390" lastFinishedPulling="2026-01-24 00:20:33.952577968 +0000 UTC m=+1017.982548969" observedRunningTime="2026-01-24 00:21:16.585721926 +0000 UTC m=+1060.615692927" watchObservedRunningTime="2026-01-24 00:21:16.589921505 +0000 UTC m=+1060.619892496" Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.615282 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-kbtsc" podStartSLOduration=2.19904079 podStartE2EDuration="17.615264721s" podCreationTimestamp="2026-01-24 00:20:59 +0000 UTC" firstStartedPulling="2026-01-24 00:21:00.325442918 +0000 UTC m=+1044.355413919" lastFinishedPulling="2026-01-24 00:21:15.741666849 +0000 UTC m=+1059.771637850" observedRunningTime="2026-01-24 00:21:16.605745779 +0000 UTC m=+1060.635716780" watchObservedRunningTime="2026-01-24 00:21:16.615264721 +0000 UTC m=+1060.645235712" Jan 24 00:21:16 crc kubenswrapper[4676]: I0124 00:21:16.641156 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=45.838200991 podStartE2EDuration="1m22.641139204s" podCreationTimestamp="2026-01-24 00:19:54 +0000 UTC" firstStartedPulling="2026-01-24 00:19:56.637072479 +0000 UTC m=+980.667043480" lastFinishedPulling="2026-01-24 00:20:33.440010672 +0000 UTC m=+1017.469981693" observedRunningTime="2026-01-24 00:21:16.640527535 +0000 UTC m=+1060.670498536" watchObservedRunningTime="2026-01-24 00:21:16.641139204 +0000 UTC m=+1060.671110205" Jan 24 00:21:17 crc kubenswrapper[4676]: I0124 00:21:17.570667 4676 generic.go:334] "Generic (PLEG): container finished" podID="f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" containerID="f59e9569d48da0357305a5589f2cf90a01ccd73d4a805dd25787bb7bb6ebfba6" exitCode=0 Jan 24 00:21:17 crc kubenswrapper[4676]: I0124 00:21:17.570775 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jsd4n-config-lcgr8" event={"ID":"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435","Type":"ContainerDied","Data":"f59e9569d48da0357305a5589f2cf90a01ccd73d4a805dd25787bb7bb6ebfba6"} Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.110592 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.180976 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20421570-a66e-4903-937f-af2ab847b5f1-operator-scripts\") pod \"20421570-a66e-4903-937f-af2ab847b5f1\" (UID: \"20421570-a66e-4903-937f-af2ab847b5f1\") " Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.181405 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7wkn\" (UniqueName: \"kubernetes.io/projected/20421570-a66e-4903-937f-af2ab847b5f1-kube-api-access-b7wkn\") pod \"20421570-a66e-4903-937f-af2ab847b5f1\" (UID: \"20421570-a66e-4903-937f-af2ab847b5f1\") " Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.181750 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20421570-a66e-4903-937f-af2ab847b5f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20421570-a66e-4903-937f-af2ab847b5f1" (UID: "20421570-a66e-4903-937f-af2ab847b5f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.188538 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20421570-a66e-4903-937f-af2ab847b5f1-kube-api-access-b7wkn" (OuterVolumeSpecName: "kube-api-access-b7wkn") pod "20421570-a66e-4903-937f-af2ab847b5f1" (UID: "20421570-a66e-4903-937f-af2ab847b5f1"). InnerVolumeSpecName "kube-api-access-b7wkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.283537 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20421570-a66e-4903-937f-af2ab847b5f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.283573 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7wkn\" (UniqueName: \"kubernetes.io/projected/20421570-a66e-4903-937f-af2ab847b5f1-kube-api-access-b7wkn\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.579938 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drjg7" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.579933 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drjg7" event={"ID":"20421570-a66e-4903-937f-af2ab847b5f1","Type":"ContainerDied","Data":"6fe7e9ee134a3384cdb14a9c06f284d65f4ca15260ccd9bddf8bd062136f714f"} Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.580066 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fe7e9ee134a3384cdb14a9c06f284d65f4ca15260ccd9bddf8bd062136f714f" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.582225 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"52b0c7a564ec2421ae93c6f06e2357b89add4953560b35b93460705f5e691400"} Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.582260 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"6792f08274ea491e2f47a7a92b9b8f41450cfcd3cd3fd728fc94129facdea2ae"} Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.582275 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"f8580f6d6428314292177ea4a66fb98598d83fc57cbb9a1593ab0fae2d6f4a0f"} Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.857467 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.889657 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-log-ovn\") pod \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.889709 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8s74\" (UniqueName: \"kubernetes.io/projected/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-kube-api-access-w8s74\") pod \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.889801 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run\") pod \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.889881 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-additional-scripts\") pod \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.889980 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-scripts\") pod \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.890007 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run-ovn\") pod \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\" (UID: \"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435\") " Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.890361 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" (UID: "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.890409 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" (UID: "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.891324 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run" (OuterVolumeSpecName: "var-run") pod "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" (UID: "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.891859 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" (UID: "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.892279 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-scripts" (OuterVolumeSpecName: "scripts") pod "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" (UID: "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.893590 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-kube-api-access-w8s74" (OuterVolumeSpecName: "kube-api-access-w8s74") pod "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" (UID: "f5533b3f-acbd-48c8-ad6f-96ba8ea7a435"). InnerVolumeSpecName "kube-api-access-w8s74". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.991671 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.991704 4676 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.991717 4676 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.991730 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8s74\" (UniqueName: \"kubernetes.io/projected/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-kube-api-access-w8s74\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.991746 4676 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-var-run\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:18 crc kubenswrapper[4676]: I0124 00:21:18.991759 4676 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:19 crc kubenswrapper[4676]: I0124 00:21:19.591964 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jsd4n-config-lcgr8" Jan 24 00:21:19 crc kubenswrapper[4676]: I0124 00:21:19.591959 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jsd4n-config-lcgr8" event={"ID":"f5533b3f-acbd-48c8-ad6f-96ba8ea7a435","Type":"ContainerDied","Data":"6ab4ddbfca70e6160be9f44ee5e62ac3fe6c06d281ba30970058467edbdbafdb"} Jan 24 00:21:19 crc kubenswrapper[4676]: I0124 00:21:19.592298 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ab4ddbfca70e6160be9f44ee5e62ac3fe6c06d281ba30970058467edbdbafdb" Jan 24 00:21:19 crc kubenswrapper[4676]: I0124 00:21:19.595540 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"1edba0950f06bc40eef9983d691b1b32ddfeb63659c38025dfd483b9df5ccb22"} Jan 24 00:21:20 crc kubenswrapper[4676]: I0124 00:21:20.099283 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jsd4n-config-lcgr8"] Jan 24 00:21:20 crc kubenswrapper[4676]: I0124 00:21:20.112473 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jsd4n-config-lcgr8"] Jan 24 00:21:20 crc kubenswrapper[4676]: I0124 00:21:20.263605 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" path="/var/lib/kubelet/pods/f5533b3f-acbd-48c8-ad6f-96ba8ea7a435/volumes" Jan 24 00:21:20 crc kubenswrapper[4676]: I0124 00:21:20.371989 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jsd4n" Jan 24 00:21:21 crc kubenswrapper[4676]: I0124 00:21:21.609805 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"3aecc2f3ca1fbd1049f9006533fdf5d01f6bf942c5ee181d9c4658231ef5e7fc"} Jan 24 00:21:21 crc kubenswrapper[4676]: I0124 00:21:21.610078 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"150a2912253199f68a039847a72a77917d84e33ab2aefe72f6d106206c8e35f2"} Jan 24 00:21:22 crc kubenswrapper[4676]: I0124 00:21:22.623077 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"c9deb54ab3dcda13383c05b900b169d3fe544744cca71b059f30e3d3261057cb"} Jan 24 00:21:22 crc kubenswrapper[4676]: I0124 00:21:22.623414 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"b3e61ed8108d731569f941607d6be9445778ccdb545b6016a97194e0815a77bc"} Jan 24 00:21:23 crc kubenswrapper[4676]: I0124 00:21:23.659690 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"005807bdc57d3d5ebc8b98f4b159a6dad3b74286736b8b7a003461a08a73aecd"} Jan 24 00:21:23 crc kubenswrapper[4676]: I0124 00:21:23.659934 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"30bc488fc17539b0bc9a0c5425e8464ec0588d5aa449a80877b372000a22b666"} Jan 24 00:21:23 crc kubenswrapper[4676]: I0124 00:21:23.659944 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"696beba936985a72ca3611ac139646e8cca505fa8af18342f82ab73e4e8837e0"} Jan 24 00:21:23 crc kubenswrapper[4676]: I0124 00:21:23.659953 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"a01bdae7755588308544f3dcea427b17fe2699df1d983433d0384a78222ddb5f"} Jan 24 00:21:24 crc kubenswrapper[4676]: I0124 00:21:24.676796 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"79b8e0c78855d2388c4c173e7757e18b5c4325314488e65f7aa9ec8b93c76962"} Jan 24 00:21:24 crc kubenswrapper[4676]: I0124 00:21:24.677210 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"a18bffdbd875d6689e14104137855c1f19cd5172e09b69e38f77e3612a7ce891"} Jan 24 00:21:24 crc kubenswrapper[4676]: I0124 00:21:24.677228 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4620e725-9218-461b-a56d-104bcb7f1df4","Type":"ContainerStarted","Data":"4687615383f49d79f3e02c88f15e42314132212d09bb3a522db9cccd1d403781"} Jan 24 00:21:24 crc kubenswrapper[4676]: I0124 00:21:24.713780 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=38.074208936 podStartE2EDuration="44.713765816s" podCreationTimestamp="2026-01-24 00:20:40 +0000 UTC" firstStartedPulling="2026-01-24 00:21:16.34029028 +0000 UTC m=+1060.370261281" lastFinishedPulling="2026-01-24 00:21:22.97984716 +0000 UTC m=+1067.009818161" observedRunningTime="2026-01-24 00:21:24.710747163 +0000 UTC m=+1068.740718164" watchObservedRunningTime="2026-01-24 00:21:24.713765816 +0000 UTC m=+1068.743736817" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.043744 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-kvxlz"] Jan 24 00:21:25 crc kubenswrapper[4676]: E0124 00:21:25.044236 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" containerName="ovn-config" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.044251 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" containerName="ovn-config" Jan 24 00:21:25 crc kubenswrapper[4676]: E0124 00:21:25.044282 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20421570-a66e-4903-937f-af2ab847b5f1" containerName="mariadb-account-create-update" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.044288 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="20421570-a66e-4903-937f-af2ab847b5f1" containerName="mariadb-account-create-update" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.044448 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="20421570-a66e-4903-937f-af2ab847b5f1" containerName="mariadb-account-create-update" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.044466 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5533b3f-acbd-48c8-ad6f-96ba8ea7a435" containerName="ovn-config" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.045181 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.047263 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.067122 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-kvxlz"] Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.138473 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.138523 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.138543 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxwv2\" (UniqueName: \"kubernetes.io/projected/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-kube-api-access-hxwv2\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.138577 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-config\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.138604 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.138650 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.239940 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.239986 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxwv2\" (UniqueName: \"kubernetes.io/projected/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-kube-api-access-hxwv2\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.240024 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-config\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.240072 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.240857 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.241020 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-config\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.241041 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.240106 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.241177 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.241184 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.242173 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.257763 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxwv2\" (UniqueName: \"kubernetes.io/projected/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-kube-api-access-hxwv2\") pod \"dnsmasq-dns-5c79d794d7-kvxlz\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.360484 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.803408 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Jan 24 00:21:25 crc kubenswrapper[4676]: I0124 00:21:25.842790 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-kvxlz"] Jan 24 00:21:26 crc kubenswrapper[4676]: I0124 00:21:26.129510 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:21:26 crc kubenswrapper[4676]: I0124 00:21:26.690938 4676 generic.go:334] "Generic (PLEG): container finished" podID="7ebdc261-7c0d-49a2-9233-868fda906788" containerID="1c08827f5b622dc0d2e57d1709fbff0d00bb4f69592994ac4ce101a373fb04cb" exitCode=0 Jan 24 00:21:26 crc kubenswrapper[4676]: I0124 00:21:26.691005 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kbtsc" event={"ID":"7ebdc261-7c0d-49a2-9233-868fda906788","Type":"ContainerDied","Data":"1c08827f5b622dc0d2e57d1709fbff0d00bb4f69592994ac4ce101a373fb04cb"} Jan 24 00:21:26 crc kubenswrapper[4676]: I0124 00:21:26.692624 4676 generic.go:334] "Generic (PLEG): container finished" podID="916cbdfc-24b9-47e7-80a6-bb5bb409ae51" containerID="4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e" exitCode=0 Jan 24 00:21:26 crc kubenswrapper[4676]: I0124 00:21:26.692658 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" event={"ID":"916cbdfc-24b9-47e7-80a6-bb5bb409ae51","Type":"ContainerDied","Data":"4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e"} Jan 24 00:21:26 crc kubenswrapper[4676]: I0124 00:21:26.692675 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" event={"ID":"916cbdfc-24b9-47e7-80a6-bb5bb409ae51","Type":"ContainerStarted","Data":"01e3b1827353ad5bf4f84620543e191fa4286a5efa456ae8149d657a438c89ff"} Jan 24 00:21:27 crc kubenswrapper[4676]: I0124 00:21:27.703564 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" event={"ID":"916cbdfc-24b9-47e7-80a6-bb5bb409ae51","Type":"ContainerStarted","Data":"691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4"} Jan 24 00:21:27 crc kubenswrapper[4676]: I0124 00:21:27.703858 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:27 crc kubenswrapper[4676]: I0124 00:21:27.728909 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" podStartSLOduration=2.728893347 podStartE2EDuration="2.728893347s" podCreationTimestamp="2026-01-24 00:21:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:27.727492164 +0000 UTC m=+1071.757463165" watchObservedRunningTime="2026-01-24 00:21:27.728893347 +0000 UTC m=+1071.758864348" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.101147 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kbtsc" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.192218 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-combined-ca-bundle\") pod \"7ebdc261-7c0d-49a2-9233-868fda906788\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.192275 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-config-data\") pod \"7ebdc261-7c0d-49a2-9233-868fda906788\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.192317 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-db-sync-config-data\") pod \"7ebdc261-7c0d-49a2-9233-868fda906788\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.192368 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6qmf\" (UniqueName: \"kubernetes.io/projected/7ebdc261-7c0d-49a2-9233-868fda906788-kube-api-access-t6qmf\") pod \"7ebdc261-7c0d-49a2-9233-868fda906788\" (UID: \"7ebdc261-7c0d-49a2-9233-868fda906788\") " Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.197632 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ebdc261-7c0d-49a2-9233-868fda906788-kube-api-access-t6qmf" (OuterVolumeSpecName: "kube-api-access-t6qmf") pod "7ebdc261-7c0d-49a2-9233-868fda906788" (UID: "7ebdc261-7c0d-49a2-9233-868fda906788"). InnerVolumeSpecName "kube-api-access-t6qmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.205509 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "7ebdc261-7c0d-49a2-9233-868fda906788" (UID: "7ebdc261-7c0d-49a2-9233-868fda906788"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.216564 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ebdc261-7c0d-49a2-9233-868fda906788" (UID: "7ebdc261-7c0d-49a2-9233-868fda906788"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.233941 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-config-data" (OuterVolumeSpecName: "config-data") pod "7ebdc261-7c0d-49a2-9233-868fda906788" (UID: "7ebdc261-7c0d-49a2-9233-868fda906788"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.295196 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6qmf\" (UniqueName: \"kubernetes.io/projected/7ebdc261-7c0d-49a2-9233-868fda906788-kube-api-access-t6qmf\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.295226 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.295236 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.295246 4676 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/7ebdc261-7c0d-49a2-9233-868fda906788-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.711467 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-kbtsc" event={"ID":"7ebdc261-7c0d-49a2-9233-868fda906788","Type":"ContainerDied","Data":"e532ae412cda92f465aa9e050762bb063a13c20d49e7ab63fe45cd1ee4ddc01b"} Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.711510 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e532ae412cda92f465aa9e050762bb063a13c20d49e7ab63fe45cd1ee4ddc01b" Jan 24 00:21:28 crc kubenswrapper[4676]: I0124 00:21:28.711516 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-kbtsc" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.182507 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-kvxlz"] Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.222994 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-p2f7j"] Jan 24 00:21:29 crc kubenswrapper[4676]: E0124 00:21:29.223288 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ebdc261-7c0d-49a2-9233-868fda906788" containerName="glance-db-sync" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.223303 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ebdc261-7c0d-49a2-9233-868fda906788" containerName="glance-db-sync" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.223461 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ebdc261-7c0d-49a2-9233-868fda906788" containerName="glance-db-sync" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.224187 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.248551 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-p2f7j"] Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.309854 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.309896 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.309942 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.310104 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.310273 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68mnw\" (UniqueName: \"kubernetes.io/projected/92c7f906-f96b-4d55-b771-6d6425b32b85-kube-api-access-68mnw\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.310453 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-config\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.412057 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68mnw\" (UniqueName: \"kubernetes.io/projected/92c7f906-f96b-4d55-b771-6d6425b32b85-kube-api-access-68mnw\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.412127 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-config\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.412195 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.412216 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.412245 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.412277 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.413089 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.414061 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-config\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.414644 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.415185 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.415742 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.431532 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68mnw\" (UniqueName: \"kubernetes.io/projected/92c7f906-f96b-4d55-b771-6d6425b32b85-kube-api-access-68mnw\") pod \"dnsmasq-dns-5f59b8f679-p2f7j\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.542184 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:29 crc kubenswrapper[4676]: I0124 00:21:29.718520 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" podUID="916cbdfc-24b9-47e7-80a6-bb5bb409ae51" containerName="dnsmasq-dns" containerID="cri-o://691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4" gracePeriod=10 Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.053956 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-p2f7j"] Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.209657 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.327234 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-sb\") pod \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.327579 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxwv2\" (UniqueName: \"kubernetes.io/projected/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-kube-api-access-hxwv2\") pod \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.327636 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-config\") pod \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.327681 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-svc\") pod \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.327727 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-swift-storage-0\") pod \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.327762 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-nb\") pod \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\" (UID: \"916cbdfc-24b9-47e7-80a6-bb5bb409ae51\") " Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.334091 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-kube-api-access-hxwv2" (OuterVolumeSpecName: "kube-api-access-hxwv2") pod "916cbdfc-24b9-47e7-80a6-bb5bb409ae51" (UID: "916cbdfc-24b9-47e7-80a6-bb5bb409ae51"). InnerVolumeSpecName "kube-api-access-hxwv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.391635 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "916cbdfc-24b9-47e7-80a6-bb5bb409ae51" (UID: "916cbdfc-24b9-47e7-80a6-bb5bb409ae51"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.392032 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "916cbdfc-24b9-47e7-80a6-bb5bb409ae51" (UID: "916cbdfc-24b9-47e7-80a6-bb5bb409ae51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.396561 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "916cbdfc-24b9-47e7-80a6-bb5bb409ae51" (UID: "916cbdfc-24b9-47e7-80a6-bb5bb409ae51"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.414180 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-config" (OuterVolumeSpecName: "config") pod "916cbdfc-24b9-47e7-80a6-bb5bb409ae51" (UID: "916cbdfc-24b9-47e7-80a6-bb5bb409ae51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.417386 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "916cbdfc-24b9-47e7-80a6-bb5bb409ae51" (UID: "916cbdfc-24b9-47e7-80a6-bb5bb409ae51"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.430780 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.430809 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.430821 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.430832 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxwv2\" (UniqueName: \"kubernetes.io/projected/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-kube-api-access-hxwv2\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.430841 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.430850 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/916cbdfc-24b9-47e7-80a6-bb5bb409ae51-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.754741 4676 generic.go:334] "Generic (PLEG): container finished" podID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerID="041801085cb8948d9dc29d8e4969e7d85a03615218568807e333c3bd95e75583" exitCode=0 Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.755630 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" event={"ID":"92c7f906-f96b-4d55-b771-6d6425b32b85","Type":"ContainerDied","Data":"041801085cb8948d9dc29d8e4969e7d85a03615218568807e333c3bd95e75583"} Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.755712 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" event={"ID":"92c7f906-f96b-4d55-b771-6d6425b32b85","Type":"ContainerStarted","Data":"af5442d400f3ec8197d8bc02bc11e29bed515b4e2a1c69855707d8adbf0ab335"} Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.760020 4676 generic.go:334] "Generic (PLEG): container finished" podID="916cbdfc-24b9-47e7-80a6-bb5bb409ae51" containerID="691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4" exitCode=0 Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.760055 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" event={"ID":"916cbdfc-24b9-47e7-80a6-bb5bb409ae51","Type":"ContainerDied","Data":"691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4"} Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.760082 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" event={"ID":"916cbdfc-24b9-47e7-80a6-bb5bb409ae51","Type":"ContainerDied","Data":"01e3b1827353ad5bf4f84620543e191fa4286a5efa456ae8149d657a438c89ff"} Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.760103 4676 scope.go:117] "RemoveContainer" containerID="691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.760210 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-kvxlz" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.911777 4676 scope.go:117] "RemoveContainer" containerID="4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.931687 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-kvxlz"] Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.937631 4676 scope.go:117] "RemoveContainer" containerID="691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4" Jan 24 00:21:30 crc kubenswrapper[4676]: E0124 00:21:30.938154 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4\": container with ID starting with 691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4 not found: ID does not exist" containerID="691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.938183 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4"} err="failed to get container status \"691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4\": rpc error: code = NotFound desc = could not find container \"691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4\": container with ID starting with 691040f96d1b5ce6cc15046d00078df9b6ee0a523e5850e6110844a0d56b04d4 not found: ID does not exist" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.938204 4676 scope.go:117] "RemoveContainer" containerID="4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e" Jan 24 00:21:30 crc kubenswrapper[4676]: E0124 00:21:30.938566 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e\": container with ID starting with 4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e not found: ID does not exist" containerID="4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.938607 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e"} err="failed to get container status \"4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e\": rpc error: code = NotFound desc = could not find container \"4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e\": container with ID starting with 4146a6e7b1d24d3fef677ddf302a756b49304320ffdd75256ecd3013d9af408e not found: ID does not exist" Jan 24 00:21:30 crc kubenswrapper[4676]: I0124 00:21:30.940313 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-kvxlz"] Jan 24 00:21:31 crc kubenswrapper[4676]: I0124 00:21:31.777061 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" event={"ID":"92c7f906-f96b-4d55-b771-6d6425b32b85","Type":"ContainerStarted","Data":"eecf36d183fa31e03de5954a2aec194cfc8588694d23fe41c8598d42bf4437a3"} Jan 24 00:21:31 crc kubenswrapper[4676]: I0124 00:21:31.777667 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:31 crc kubenswrapper[4676]: I0124 00:21:31.811796 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" podStartSLOduration=2.811772815 podStartE2EDuration="2.811772815s" podCreationTimestamp="2026-01-24 00:21:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:31.80867756 +0000 UTC m=+1075.838648571" watchObservedRunningTime="2026-01-24 00:21:31.811772815 +0000 UTC m=+1075.841743856" Jan 24 00:21:32 crc kubenswrapper[4676]: I0124 00:21:32.272278 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="916cbdfc-24b9-47e7-80a6-bb5bb409ae51" path="/var/lib/kubelet/pods/916cbdfc-24b9-47e7-80a6-bb5bb409ae51/volumes" Jan 24 00:21:35 crc kubenswrapper[4676]: I0124 00:21:35.806579 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.121203 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-bcttg"] Jan 24 00:21:36 crc kubenswrapper[4676]: E0124 00:21:36.121738 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="916cbdfc-24b9-47e7-80a6-bb5bb409ae51" containerName="dnsmasq-dns" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.121753 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="916cbdfc-24b9-47e7-80a6-bb5bb409ae51" containerName="dnsmasq-dns" Jan 24 00:21:36 crc kubenswrapper[4676]: E0124 00:21:36.121783 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="916cbdfc-24b9-47e7-80a6-bb5bb409ae51" containerName="init" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.121790 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="916cbdfc-24b9-47e7-80a6-bb5bb409ae51" containerName="init" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.121930 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="916cbdfc-24b9-47e7-80a6-bb5bb409ae51" containerName="dnsmasq-dns" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.122389 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.136329 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bcttg"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.240322 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-b4lc5"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.241407 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.252946 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d058-account-create-update-pjzm7"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.253951 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.256073 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.283207 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d058-account-create-update-pjzm7"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.283232 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-b4lc5"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.288536 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b488c3f8-bc04-4c8d-98b7-72145ae9e948-operator-scripts\") pod \"cinder-db-create-bcttg\" (UID: \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\") " pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.288667 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mnv8\" (UniqueName: \"kubernetes.io/projected/b488c3f8-bc04-4c8d-98b7-72145ae9e948-kube-api-access-4mnv8\") pod \"cinder-db-create-bcttg\" (UID: \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\") " pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.390365 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mnv8\" (UniqueName: \"kubernetes.io/projected/b488c3f8-bc04-4c8d-98b7-72145ae9e948-kube-api-access-4mnv8\") pod \"cinder-db-create-bcttg\" (UID: \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\") " pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.390432 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwnv8\" (UniqueName: \"kubernetes.io/projected/64a07672-60fb-4935-83e2-99f39e15427f-kube-api-access-qwnv8\") pod \"cinder-d058-account-create-update-pjzm7\" (UID: \"64a07672-60fb-4935-83e2-99f39e15427f\") " pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.390459 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-operator-scripts\") pod \"barbican-db-create-b4lc5\" (UID: \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\") " pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.390487 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmv6c\" (UniqueName: \"kubernetes.io/projected/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-kube-api-access-vmv6c\") pod \"barbican-db-create-b4lc5\" (UID: \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\") " pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.390550 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b488c3f8-bc04-4c8d-98b7-72145ae9e948-operator-scripts\") pod \"cinder-db-create-bcttg\" (UID: \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\") " pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.390566 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64a07672-60fb-4935-83e2-99f39e15427f-operator-scripts\") pod \"cinder-d058-account-create-update-pjzm7\" (UID: \"64a07672-60fb-4935-83e2-99f39e15427f\") " pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.391790 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b488c3f8-bc04-4c8d-98b7-72145ae9e948-operator-scripts\") pod \"cinder-db-create-bcttg\" (UID: \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\") " pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.417716 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mnv8\" (UniqueName: \"kubernetes.io/projected/b488c3f8-bc04-4c8d-98b7-72145ae9e948-kube-api-access-4mnv8\") pod \"cinder-db-create-bcttg\" (UID: \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\") " pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.440308 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-mtdlz"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.441489 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.444476 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.471658 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mtdlz"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.494166 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwnv8\" (UniqueName: \"kubernetes.io/projected/64a07672-60fb-4935-83e2-99f39e15427f-kube-api-access-qwnv8\") pod \"cinder-d058-account-create-update-pjzm7\" (UID: \"64a07672-60fb-4935-83e2-99f39e15427f\") " pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.494228 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-operator-scripts\") pod \"barbican-db-create-b4lc5\" (UID: \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\") " pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.494271 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmv6c\" (UniqueName: \"kubernetes.io/projected/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-kube-api-access-vmv6c\") pod \"barbican-db-create-b4lc5\" (UID: \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\") " pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.494329 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64a07672-60fb-4935-83e2-99f39e15427f-operator-scripts\") pod \"cinder-d058-account-create-update-pjzm7\" (UID: \"64a07672-60fb-4935-83e2-99f39e15427f\") " pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.495178 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64a07672-60fb-4935-83e2-99f39e15427f-operator-scripts\") pod \"cinder-d058-account-create-update-pjzm7\" (UID: \"64a07672-60fb-4935-83e2-99f39e15427f\") " pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.495944 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-operator-scripts\") pod \"barbican-db-create-b4lc5\" (UID: \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\") " pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.521222 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwnv8\" (UniqueName: \"kubernetes.io/projected/64a07672-60fb-4935-83e2-99f39e15427f-kube-api-access-qwnv8\") pod \"cinder-d058-account-create-update-pjzm7\" (UID: \"64a07672-60fb-4935-83e2-99f39e15427f\") " pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.539904 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-4766-account-create-update-bbg95"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.540815 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.547102 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.549004 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmv6c\" (UniqueName: \"kubernetes.io/projected/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-kube-api-access-vmv6c\") pod \"barbican-db-create-b4lc5\" (UID: \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\") " pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.570267 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.572260 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4766-account-create-update-bbg95"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.591008 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.597459 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4fb3ef-0cde-499c-8018-14cc96b495f5-operator-scripts\") pod \"neutron-db-create-mtdlz\" (UID: \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\") " pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.597515 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw24b\" (UniqueName: \"kubernetes.io/projected/2d4fb3ef-0cde-499c-8018-14cc96b495f5-kube-api-access-zw24b\") pod \"neutron-db-create-mtdlz\" (UID: \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\") " pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.679307 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wg6tc"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.684972 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wg6tc"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.685083 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.688737 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fcgg9" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.688915 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.691766 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.695072 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.701418 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-operator-scripts\") pod \"barbican-4766-account-create-update-bbg95\" (UID: \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\") " pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.701460 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4fb3ef-0cde-499c-8018-14cc96b495f5-operator-scripts\") pod \"neutron-db-create-mtdlz\" (UID: \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\") " pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.701500 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw24b\" (UniqueName: \"kubernetes.io/projected/2d4fb3ef-0cde-499c-8018-14cc96b495f5-kube-api-access-zw24b\") pod \"neutron-db-create-mtdlz\" (UID: \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\") " pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.701550 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stcpz\" (UniqueName: \"kubernetes.io/projected/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-kube-api-access-stcpz\") pod \"barbican-4766-account-create-update-bbg95\" (UID: \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\") " pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.710684 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4fb3ef-0cde-499c-8018-14cc96b495f5-operator-scripts\") pod \"neutron-db-create-mtdlz\" (UID: \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\") " pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.726834 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw24b\" (UniqueName: \"kubernetes.io/projected/2d4fb3ef-0cde-499c-8018-14cc96b495f5-kube-api-access-zw24b\") pod \"neutron-db-create-mtdlz\" (UID: \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\") " pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.745975 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-ba07-account-create-update-hw2sx"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.746908 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.761424 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.771206 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ba07-account-create-update-hw2sx"] Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.805556 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhdgs\" (UniqueName: \"kubernetes.io/projected/30027bc7-2637-4abb-9568-1190cf80a3a5-kube-api-access-bhdgs\") pod \"keystone-db-sync-wg6tc\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.805632 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stcpz\" (UniqueName: \"kubernetes.io/projected/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-kube-api-access-stcpz\") pod \"barbican-4766-account-create-update-bbg95\" (UID: \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\") " pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.805686 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ace7871-3944-4c72-980f-c9d5e7d65c71-operator-scripts\") pod \"neutron-ba07-account-create-update-hw2sx\" (UID: \"9ace7871-3944-4c72-980f-c9d5e7d65c71\") " pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.805733 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-config-data\") pod \"keystone-db-sync-wg6tc\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.805759 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-combined-ca-bundle\") pod \"keystone-db-sync-wg6tc\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.805779 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb6ck\" (UniqueName: \"kubernetes.io/projected/9ace7871-3944-4c72-980f-c9d5e7d65c71-kube-api-access-qb6ck\") pod \"neutron-ba07-account-create-update-hw2sx\" (UID: \"9ace7871-3944-4c72-980f-c9d5e7d65c71\") " pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.805808 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-operator-scripts\") pod \"barbican-4766-account-create-update-bbg95\" (UID: \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\") " pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.806470 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-operator-scripts\") pod \"barbican-4766-account-create-update-bbg95\" (UID: \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\") " pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.836891 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stcpz\" (UniqueName: \"kubernetes.io/projected/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-kube-api-access-stcpz\") pod \"barbican-4766-account-create-update-bbg95\" (UID: \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\") " pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.901177 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.908024 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-config-data\") pod \"keystone-db-sync-wg6tc\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.908070 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-combined-ca-bundle\") pod \"keystone-db-sync-wg6tc\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.908113 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb6ck\" (UniqueName: \"kubernetes.io/projected/9ace7871-3944-4c72-980f-c9d5e7d65c71-kube-api-access-qb6ck\") pod \"neutron-ba07-account-create-update-hw2sx\" (UID: \"9ace7871-3944-4c72-980f-c9d5e7d65c71\") " pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.908162 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhdgs\" (UniqueName: \"kubernetes.io/projected/30027bc7-2637-4abb-9568-1190cf80a3a5-kube-api-access-bhdgs\") pod \"keystone-db-sync-wg6tc\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.908214 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ace7871-3944-4c72-980f-c9d5e7d65c71-operator-scripts\") pod \"neutron-ba07-account-create-update-hw2sx\" (UID: \"9ace7871-3944-4c72-980f-c9d5e7d65c71\") " pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.908811 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ace7871-3944-4c72-980f-c9d5e7d65c71-operator-scripts\") pod \"neutron-ba07-account-create-update-hw2sx\" (UID: \"9ace7871-3944-4c72-980f-c9d5e7d65c71\") " pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.913280 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-combined-ca-bundle\") pod \"keystone-db-sync-wg6tc\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.913306 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-config-data\") pod \"keystone-db-sync-wg6tc\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.941322 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhdgs\" (UniqueName: \"kubernetes.io/projected/30027bc7-2637-4abb-9568-1190cf80a3a5-kube-api-access-bhdgs\") pod \"keystone-db-sync-wg6tc\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.952737 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb6ck\" (UniqueName: \"kubernetes.io/projected/9ace7871-3944-4c72-980f-c9d5e7d65c71-kube-api-access-qb6ck\") pod \"neutron-ba07-account-create-update-hw2sx\" (UID: \"9ace7871-3944-4c72-980f-c9d5e7d65c71\") " pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:36 crc kubenswrapper[4676]: I0124 00:21:36.996902 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.015514 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.058896 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-bcttg"] Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.207790 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.393362 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d058-account-create-update-pjzm7"] Jan 24 00:21:37 crc kubenswrapper[4676]: W0124 00:21:37.406942 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64a07672_60fb_4935_83e2_99f39e15427f.slice/crio-aa778d2b9641cb4a032b28b9537d81b843cc0344050ae97830c3a489d9508916 WatchSource:0}: Error finding container aa778d2b9641cb4a032b28b9537d81b843cc0344050ae97830c3a489d9508916: Status 404 returned error can't find the container with id aa778d2b9641cb4a032b28b9537d81b843cc0344050ae97830c3a489d9508916 Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.420367 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-b4lc5"] Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.506302 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4766-account-create-update-bbg95"] Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.561906 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wg6tc"] Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.580587 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mtdlz"] Jan 24 00:21:37 crc kubenswrapper[4676]: W0124 00:21:37.596580 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30027bc7_2637_4abb_9568_1190cf80a3a5.slice/crio-3b055c8876511446c351694e38bb4288f579441b122767bc434100cd495602c9 WatchSource:0}: Error finding container 3b055c8876511446c351694e38bb4288f579441b122767bc434100cd495602c9: Status 404 returned error can't find the container with id 3b055c8876511446c351694e38bb4288f579441b122767bc434100cd495602c9 Jan 24 00:21:37 crc kubenswrapper[4676]: W0124 00:21:37.611922 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d4fb3ef_0cde_499c_8018_14cc96b495f5.slice/crio-ca93660ad2310b325562d8e1908e400fca734f442cb3fbc769a5593ef16a799a WatchSource:0}: Error finding container ca93660ad2310b325562d8e1908e400fca734f442cb3fbc769a5593ef16a799a: Status 404 returned error can't find the container with id ca93660ad2310b325562d8e1908e400fca734f442cb3fbc769a5593ef16a799a Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.836063 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wg6tc" event={"ID":"30027bc7-2637-4abb-9568-1190cf80a3a5","Type":"ContainerStarted","Data":"3b055c8876511446c351694e38bb4288f579441b122767bc434100cd495602c9"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.838585 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4766-account-create-update-bbg95" event={"ID":"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09","Type":"ContainerStarted","Data":"c0f6b7c08c73c5cd3ba2389c70cde6261bfdc95261dc1f7fcb14c46b9908056d"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.838637 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4766-account-create-update-bbg95" event={"ID":"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09","Type":"ContainerStarted","Data":"497443544a8b1e5d2264898469d70d9d8eb5bc684f16d3e8ec4d93ed1fa60146"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.841642 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mtdlz" event={"ID":"2d4fb3ef-0cde-499c-8018-14cc96b495f5","Type":"ContainerStarted","Data":"b062d260d4a7b754ef8f1768435c8c0945467373bcd6d26d2a060794a30067d3"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.841673 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mtdlz" event={"ID":"2d4fb3ef-0cde-499c-8018-14cc96b495f5","Type":"ContainerStarted","Data":"ca93660ad2310b325562d8e1908e400fca734f442cb3fbc769a5593ef16a799a"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.846385 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d058-account-create-update-pjzm7" event={"ID":"64a07672-60fb-4935-83e2-99f39e15427f","Type":"ContainerStarted","Data":"acb8fc5743669070f34cea9944d1cdf719e2212dfd00759a3d455ff74f1aa3b7"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.846446 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d058-account-create-update-pjzm7" event={"ID":"64a07672-60fb-4935-83e2-99f39e15427f","Type":"ContainerStarted","Data":"aa778d2b9641cb4a032b28b9537d81b843cc0344050ae97830c3a489d9508916"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.848132 4676 generic.go:334] "Generic (PLEG): container finished" podID="b488c3f8-bc04-4c8d-98b7-72145ae9e948" containerID="89ec371006cc8981cf4d3c50f26065f195057a546aba2faa19252efcd286b309" exitCode=0 Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.848203 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bcttg" event={"ID":"b488c3f8-bc04-4c8d-98b7-72145ae9e948","Type":"ContainerDied","Data":"89ec371006cc8981cf4d3c50f26065f195057a546aba2faa19252efcd286b309"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.848354 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bcttg" event={"ID":"b488c3f8-bc04-4c8d-98b7-72145ae9e948","Type":"ContainerStarted","Data":"96d1f38596ab4c33a9c07bd749a1c1fae3129b0d58a93768d2af5c2ff4e39a89"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.850139 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b4lc5" event={"ID":"64efcf33-9ffe-402d-b0b0-9cf53c7a5495","Type":"ContainerStarted","Data":"bfee0ea90c79ca47b4f4a9473a613bbee236021e6a49db23a4490b8cde0e9a39"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.850222 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b4lc5" event={"ID":"64efcf33-9ffe-402d-b0b0-9cf53c7a5495","Type":"ContainerStarted","Data":"e89a4d04ab1c50604d49d0624984dbc3a381642b8ef5d82c2bf3da9607b0c956"} Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.854670 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-4766-account-create-update-bbg95" podStartSLOduration=1.8546523019999999 podStartE2EDuration="1.854652302s" podCreationTimestamp="2026-01-24 00:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:37.851137965 +0000 UTC m=+1081.881108966" watchObservedRunningTime="2026-01-24 00:21:37.854652302 +0000 UTC m=+1081.884623303" Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.881741 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-mtdlz" podStartSLOduration=1.881722911 podStartE2EDuration="1.881722911s" podCreationTimestamp="2026-01-24 00:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:37.876966175 +0000 UTC m=+1081.906937166" watchObservedRunningTime="2026-01-24 00:21:37.881722911 +0000 UTC m=+1081.911693912" Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.902559 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-b4lc5" podStartSLOduration=1.9025422779999999 podStartE2EDuration="1.902542278s" podCreationTimestamp="2026-01-24 00:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:37.896237805 +0000 UTC m=+1081.926208806" watchObservedRunningTime="2026-01-24 00:21:37.902542278 +0000 UTC m=+1081.932513279" Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.919156 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-d058-account-create-update-pjzm7" podStartSLOduration=1.919139946 podStartE2EDuration="1.919139946s" podCreationTimestamp="2026-01-24 00:21:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:37.914192505 +0000 UTC m=+1081.944163506" watchObservedRunningTime="2026-01-24 00:21:37.919139946 +0000 UTC m=+1081.949110947" Jan 24 00:21:37 crc kubenswrapper[4676]: I0124 00:21:37.933391 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-ba07-account-create-update-hw2sx"] Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.889976 4676 generic.go:334] "Generic (PLEG): container finished" podID="9ace7871-3944-4c72-980f-c9d5e7d65c71" containerID="c87e9f17f624e372e522a3229d29135c0ab3c4953ce39a4a0c79e4002fe69658" exitCode=0 Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.890335 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ba07-account-create-update-hw2sx" event={"ID":"9ace7871-3944-4c72-980f-c9d5e7d65c71","Type":"ContainerDied","Data":"c87e9f17f624e372e522a3229d29135c0ab3c4953ce39a4a0c79e4002fe69658"} Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.890362 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ba07-account-create-update-hw2sx" event={"ID":"9ace7871-3944-4c72-980f-c9d5e7d65c71","Type":"ContainerStarted","Data":"7d5bd01e9496b9422e6f2c0b4d3d95295f8f687c97766a57ad1bfe5c28074b14"} Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.891997 4676 generic.go:334] "Generic (PLEG): container finished" podID="a61f15c0-2fa1-4d69-bb80-471d6e4e7e09" containerID="c0f6b7c08c73c5cd3ba2389c70cde6261bfdc95261dc1f7fcb14c46b9908056d" exitCode=0 Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.892043 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4766-account-create-update-bbg95" event={"ID":"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09","Type":"ContainerDied","Data":"c0f6b7c08c73c5cd3ba2389c70cde6261bfdc95261dc1f7fcb14c46b9908056d"} Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.912678 4676 generic.go:334] "Generic (PLEG): container finished" podID="2d4fb3ef-0cde-499c-8018-14cc96b495f5" containerID="b062d260d4a7b754ef8f1768435c8c0945467373bcd6d26d2a060794a30067d3" exitCode=0 Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.912962 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mtdlz" event={"ID":"2d4fb3ef-0cde-499c-8018-14cc96b495f5","Type":"ContainerDied","Data":"b062d260d4a7b754ef8f1768435c8c0945467373bcd6d26d2a060794a30067d3"} Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.948075 4676 generic.go:334] "Generic (PLEG): container finished" podID="64a07672-60fb-4935-83e2-99f39e15427f" containerID="acb8fc5743669070f34cea9944d1cdf719e2212dfd00759a3d455ff74f1aa3b7" exitCode=0 Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.948177 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d058-account-create-update-pjzm7" event={"ID":"64a07672-60fb-4935-83e2-99f39e15427f","Type":"ContainerDied","Data":"acb8fc5743669070f34cea9944d1cdf719e2212dfd00759a3d455ff74f1aa3b7"} Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.963546 4676 generic.go:334] "Generic (PLEG): container finished" podID="64efcf33-9ffe-402d-b0b0-9cf53c7a5495" containerID="bfee0ea90c79ca47b4f4a9473a613bbee236021e6a49db23a4490b8cde0e9a39" exitCode=0 Jan 24 00:21:38 crc kubenswrapper[4676]: I0124 00:21:38.963900 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b4lc5" event={"ID":"64efcf33-9ffe-402d-b0b0-9cf53c7a5495","Type":"ContainerDied","Data":"bfee0ea90c79ca47b4f4a9473a613bbee236021e6a49db23a4490b8cde0e9a39"} Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.367076 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.456088 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b488c3f8-bc04-4c8d-98b7-72145ae9e948-operator-scripts\") pod \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\" (UID: \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\") " Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.456802 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b488c3f8-bc04-4c8d-98b7-72145ae9e948-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b488c3f8-bc04-4c8d-98b7-72145ae9e948" (UID: "b488c3f8-bc04-4c8d-98b7-72145ae9e948"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.457019 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mnv8\" (UniqueName: \"kubernetes.io/projected/b488c3f8-bc04-4c8d-98b7-72145ae9e948-kube-api-access-4mnv8\") pod \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\" (UID: \"b488c3f8-bc04-4c8d-98b7-72145ae9e948\") " Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.458246 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b488c3f8-bc04-4c8d-98b7-72145ae9e948-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.462585 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b488c3f8-bc04-4c8d-98b7-72145ae9e948-kube-api-access-4mnv8" (OuterVolumeSpecName: "kube-api-access-4mnv8") pod "b488c3f8-bc04-4c8d-98b7-72145ae9e948" (UID: "b488c3f8-bc04-4c8d-98b7-72145ae9e948"). InnerVolumeSpecName "kube-api-access-4mnv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.544445 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.562556 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mnv8\" (UniqueName: \"kubernetes.io/projected/b488c3f8-bc04-4c8d-98b7-72145ae9e948-kube-api-access-4mnv8\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.611066 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-88xjq"] Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.611506 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" podUID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" containerName="dnsmasq-dns" containerID="cri-o://1d6d0279f417026b6ba6bd78b2b2fd59bc265f683401ca01652ddd44d6b071a2" gracePeriod=10 Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.974029 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-bcttg" event={"ID":"b488c3f8-bc04-4c8d-98b7-72145ae9e948","Type":"ContainerDied","Data":"96d1f38596ab4c33a9c07bd749a1c1fae3129b0d58a93768d2af5c2ff4e39a89"} Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.974080 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96d1f38596ab4c33a9c07bd749a1c1fae3129b0d58a93768d2af5c2ff4e39a89" Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.974150 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-bcttg" Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.976948 4676 generic.go:334] "Generic (PLEG): container finished" podID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" containerID="1d6d0279f417026b6ba6bd78b2b2fd59bc265f683401ca01652ddd44d6b071a2" exitCode=0 Jan 24 00:21:39 crc kubenswrapper[4676]: I0124 00:21:39.977210 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" event={"ID":"1905ca79-a4c4-4286-8d88-2855e7b9ba4c","Type":"ContainerDied","Data":"1d6d0279f417026b6ba6bd78b2b2fd59bc265f683401ca01652ddd44d6b071a2"} Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.405022 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.413962 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.452476 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ace7871-3944-4c72-980f-c9d5e7d65c71-operator-scripts\") pod \"9ace7871-3944-4c72-980f-c9d5e7d65c71\" (UID: \"9ace7871-3944-4c72-980f-c9d5e7d65c71\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.452552 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb6ck\" (UniqueName: \"kubernetes.io/projected/9ace7871-3944-4c72-980f-c9d5e7d65c71-kube-api-access-qb6ck\") pod \"9ace7871-3944-4c72-980f-c9d5e7d65c71\" (UID: \"9ace7871-3944-4c72-980f-c9d5e7d65c71\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.452595 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stcpz\" (UniqueName: \"kubernetes.io/projected/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-kube-api-access-stcpz\") pod \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\" (UID: \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.452647 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-operator-scripts\") pod \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\" (UID: \"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.454160 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a61f15c0-2fa1-4d69-bb80-471d6e4e7e09" (UID: "a61f15c0-2fa1-4d69-bb80-471d6e4e7e09"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.454863 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ace7871-3944-4c72-980f-c9d5e7d65c71-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9ace7871-3944-4c72-980f-c9d5e7d65c71" (UID: "9ace7871-3944-4c72-980f-c9d5e7d65c71"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.464593 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-kube-api-access-stcpz" (OuterVolumeSpecName: "kube-api-access-stcpz") pod "a61f15c0-2fa1-4d69-bb80-471d6e4e7e09" (UID: "a61f15c0-2fa1-4d69-bb80-471d6e4e7e09"). InnerVolumeSpecName "kube-api-access-stcpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.464629 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ace7871-3944-4c72-980f-c9d5e7d65c71-kube-api-access-qb6ck" (OuterVolumeSpecName: "kube-api-access-qb6ck") pod "9ace7871-3944-4c72-980f-c9d5e7d65c71" (UID: "9ace7871-3944-4c72-980f-c9d5e7d65c71"). InnerVolumeSpecName "kube-api-access-qb6ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.469434 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.474424 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.540850 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.550918 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.556425 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.556456 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ace7871-3944-4c72-980f-c9d5e7d65c71-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.556469 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb6ck\" (UniqueName: \"kubernetes.io/projected/9ace7871-3944-4c72-980f-c9d5e7d65c71-kube-api-access-qb6ck\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.556481 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stcpz\" (UniqueName: \"kubernetes.io/projected/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09-kube-api-access-stcpz\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.657945 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4fb3ef-0cde-499c-8018-14cc96b495f5-operator-scripts\") pod \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\" (UID: \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658010 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-operator-scripts\") pod \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\" (UID: \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658042 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpslv\" (UniqueName: \"kubernetes.io/projected/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-kube-api-access-lpslv\") pod \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658081 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-dns-svc\") pod \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658117 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-nb\") pod \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658195 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-config\") pod \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658253 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmv6c\" (UniqueName: \"kubernetes.io/projected/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-kube-api-access-vmv6c\") pod \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\" (UID: \"64efcf33-9ffe-402d-b0b0-9cf53c7a5495\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658283 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwnv8\" (UniqueName: \"kubernetes.io/projected/64a07672-60fb-4935-83e2-99f39e15427f-kube-api-access-qwnv8\") pod \"64a07672-60fb-4935-83e2-99f39e15427f\" (UID: \"64a07672-60fb-4935-83e2-99f39e15427f\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658314 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64a07672-60fb-4935-83e2-99f39e15427f-operator-scripts\") pod \"64a07672-60fb-4935-83e2-99f39e15427f\" (UID: \"64a07672-60fb-4935-83e2-99f39e15427f\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658358 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw24b\" (UniqueName: \"kubernetes.io/projected/2d4fb3ef-0cde-499c-8018-14cc96b495f5-kube-api-access-zw24b\") pod \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\" (UID: \"2d4fb3ef-0cde-499c-8018-14cc96b495f5\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.658456 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-sb\") pod \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\" (UID: \"1905ca79-a4c4-4286-8d88-2855e7b9ba4c\") " Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.659082 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d4fb3ef-0cde-499c-8018-14cc96b495f5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d4fb3ef-0cde-499c-8018-14cc96b495f5" (UID: "2d4fb3ef-0cde-499c-8018-14cc96b495f5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.659431 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "64efcf33-9ffe-402d-b0b0-9cf53c7a5495" (UID: "64efcf33-9ffe-402d-b0b0-9cf53c7a5495"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.659664 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64a07672-60fb-4935-83e2-99f39e15427f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "64a07672-60fb-4935-83e2-99f39e15427f" (UID: "64a07672-60fb-4935-83e2-99f39e15427f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.662804 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d4fb3ef-0cde-499c-8018-14cc96b495f5-kube-api-access-zw24b" (OuterVolumeSpecName: "kube-api-access-zw24b") pod "2d4fb3ef-0cde-499c-8018-14cc96b495f5" (UID: "2d4fb3ef-0cde-499c-8018-14cc96b495f5"). InnerVolumeSpecName "kube-api-access-zw24b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.663880 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-kube-api-access-vmv6c" (OuterVolumeSpecName: "kube-api-access-vmv6c") pod "64efcf33-9ffe-402d-b0b0-9cf53c7a5495" (UID: "64efcf33-9ffe-402d-b0b0-9cf53c7a5495"). InnerVolumeSpecName "kube-api-access-vmv6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.665897 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-kube-api-access-lpslv" (OuterVolumeSpecName: "kube-api-access-lpslv") pod "1905ca79-a4c4-4286-8d88-2855e7b9ba4c" (UID: "1905ca79-a4c4-4286-8d88-2855e7b9ba4c"). InnerVolumeSpecName "kube-api-access-lpslv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.672482 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64a07672-60fb-4935-83e2-99f39e15427f-kube-api-access-qwnv8" (OuterVolumeSpecName: "kube-api-access-qwnv8") pod "64a07672-60fb-4935-83e2-99f39e15427f" (UID: "64a07672-60fb-4935-83e2-99f39e15427f"). InnerVolumeSpecName "kube-api-access-qwnv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.702643 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1905ca79-a4c4-4286-8d88-2855e7b9ba4c" (UID: "1905ca79-a4c4-4286-8d88-2855e7b9ba4c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.706085 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1905ca79-a4c4-4286-8d88-2855e7b9ba4c" (UID: "1905ca79-a4c4-4286-8d88-2855e7b9ba4c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.709150 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-config" (OuterVolumeSpecName: "config") pod "1905ca79-a4c4-4286-8d88-2855e7b9ba4c" (UID: "1905ca79-a4c4-4286-8d88-2855e7b9ba4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.717854 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1905ca79-a4c4-4286-8d88-2855e7b9ba4c" (UID: "1905ca79-a4c4-4286-8d88-2855e7b9ba4c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760175 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64a07672-60fb-4935-83e2-99f39e15427f-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760212 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw24b\" (UniqueName: \"kubernetes.io/projected/2d4fb3ef-0cde-499c-8018-14cc96b495f5-kube-api-access-zw24b\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760227 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760239 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d4fb3ef-0cde-499c-8018-14cc96b495f5-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760251 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760262 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpslv\" (UniqueName: \"kubernetes.io/projected/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-kube-api-access-lpslv\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760277 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760289 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760300 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1905ca79-a4c4-4286-8d88-2855e7b9ba4c-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760313 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmv6c\" (UniqueName: \"kubernetes.io/projected/64efcf33-9ffe-402d-b0b0-9cf53c7a5495-kube-api-access-vmv6c\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:44 crc kubenswrapper[4676]: I0124 00:21:44.760325 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwnv8\" (UniqueName: \"kubernetes.io/projected/64a07672-60fb-4935-83e2-99f39e15427f-kube-api-access-qwnv8\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.029876 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mtdlz" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.030088 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mtdlz" event={"ID":"2d4fb3ef-0cde-499c-8018-14cc96b495f5","Type":"ContainerDied","Data":"ca93660ad2310b325562d8e1908e400fca734f442cb3fbc769a5593ef16a799a"} Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.030333 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca93660ad2310b325562d8e1908e400fca734f442cb3fbc769a5593ef16a799a" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.032095 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d058-account-create-update-pjzm7" event={"ID":"64a07672-60fb-4935-83e2-99f39e15427f","Type":"ContainerDied","Data":"aa778d2b9641cb4a032b28b9537d81b843cc0344050ae97830c3a489d9508916"} Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.032139 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa778d2b9641cb4a032b28b9537d81b843cc0344050ae97830c3a489d9508916" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.032219 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d058-account-create-update-pjzm7" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.037015 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-ba07-account-create-update-hw2sx" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.037477 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-ba07-account-create-update-hw2sx" event={"ID":"9ace7871-3944-4c72-980f-c9d5e7d65c71","Type":"ContainerDied","Data":"7d5bd01e9496b9422e6f2c0b4d3d95295f8f687c97766a57ad1bfe5c28074b14"} Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.037600 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d5bd01e9496b9422e6f2c0b4d3d95295f8f687c97766a57ad1bfe5c28074b14" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.038976 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-b4lc5" event={"ID":"64efcf33-9ffe-402d-b0b0-9cf53c7a5495","Type":"ContainerDied","Data":"e89a4d04ab1c50604d49d0624984dbc3a381642b8ef5d82c2bf3da9607b0c956"} Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.039074 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e89a4d04ab1c50604d49d0624984dbc3a381642b8ef5d82c2bf3da9607b0c956" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.039077 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-b4lc5" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.043966 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wg6tc" event={"ID":"30027bc7-2637-4abb-9568-1190cf80a3a5","Type":"ContainerStarted","Data":"2562adc9221a7b111bd58848de15d9abcd709528ca168c846292e16b34c5154e"} Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.048882 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" event={"ID":"1905ca79-a4c4-4286-8d88-2855e7b9ba4c","Type":"ContainerDied","Data":"b390252d36e77eb3a82a5578bfec5a23efd8e93f3aaecb313b5f2fd7e776552d"} Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.048940 4676 scope.go:117] "RemoveContainer" containerID="1d6d0279f417026b6ba6bd78b2b2fd59bc265f683401ca01652ddd44d6b071a2" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.048940 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.052579 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4766-account-create-update-bbg95" event={"ID":"a61f15c0-2fa1-4d69-bb80-471d6e4e7e09","Type":"ContainerDied","Data":"497443544a8b1e5d2264898469d70d9d8eb5bc684f16d3e8ec4d93ed1fa60146"} Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.052606 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="497443544a8b1e5d2264898469d70d9d8eb5bc684f16d3e8ec4d93ed1fa60146" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.052653 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4766-account-create-update-bbg95" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.077619 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wg6tc" podStartSLOduration=2.539306867 podStartE2EDuration="9.077601665s" podCreationTimestamp="2026-01-24 00:21:36 +0000 UTC" firstStartedPulling="2026-01-24 00:21:37.664499909 +0000 UTC m=+1081.694470910" lastFinishedPulling="2026-01-24 00:21:44.202794697 +0000 UTC m=+1088.232765708" observedRunningTime="2026-01-24 00:21:45.065800863 +0000 UTC m=+1089.095771874" watchObservedRunningTime="2026-01-24 00:21:45.077601665 +0000 UTC m=+1089.107572676" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.093807 4676 scope.go:117] "RemoveContainer" containerID="837bffae0bd65cfec1fbd0e46c064b367c74a4b7d5a87b7e24ab722d34c34bd4" Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.177038 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-88xjq"] Jan 24 00:21:45 crc kubenswrapper[4676]: I0124 00:21:45.183197 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-88xjq"] Jan 24 00:21:46 crc kubenswrapper[4676]: I0124 00:21:46.173706 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-88xjq" podUID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: i/o timeout" Jan 24 00:21:46 crc kubenswrapper[4676]: I0124 00:21:46.274743 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" path="/var/lib/kubelet/pods/1905ca79-a4c4-4286-8d88-2855e7b9ba4c/volumes" Jan 24 00:21:48 crc kubenswrapper[4676]: I0124 00:21:48.084360 4676 generic.go:334] "Generic (PLEG): container finished" podID="30027bc7-2637-4abb-9568-1190cf80a3a5" containerID="2562adc9221a7b111bd58848de15d9abcd709528ca168c846292e16b34c5154e" exitCode=0 Jan 24 00:21:48 crc kubenswrapper[4676]: I0124 00:21:48.085254 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wg6tc" event={"ID":"30027bc7-2637-4abb-9568-1190cf80a3a5","Type":"ContainerDied","Data":"2562adc9221a7b111bd58848de15d9abcd709528ca168c846292e16b34c5154e"} Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.481948 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.649750 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhdgs\" (UniqueName: \"kubernetes.io/projected/30027bc7-2637-4abb-9568-1190cf80a3a5-kube-api-access-bhdgs\") pod \"30027bc7-2637-4abb-9568-1190cf80a3a5\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.650099 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-config-data\") pod \"30027bc7-2637-4abb-9568-1190cf80a3a5\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.650823 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-combined-ca-bundle\") pod \"30027bc7-2637-4abb-9568-1190cf80a3a5\" (UID: \"30027bc7-2637-4abb-9568-1190cf80a3a5\") " Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.656039 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30027bc7-2637-4abb-9568-1190cf80a3a5-kube-api-access-bhdgs" (OuterVolumeSpecName: "kube-api-access-bhdgs") pod "30027bc7-2637-4abb-9568-1190cf80a3a5" (UID: "30027bc7-2637-4abb-9568-1190cf80a3a5"). InnerVolumeSpecName "kube-api-access-bhdgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.681101 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30027bc7-2637-4abb-9568-1190cf80a3a5" (UID: "30027bc7-2637-4abb-9568-1190cf80a3a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.699062 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-config-data" (OuterVolumeSpecName: "config-data") pod "30027bc7-2637-4abb-9568-1190cf80a3a5" (UID: "30027bc7-2637-4abb-9568-1190cf80a3a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.751937 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhdgs\" (UniqueName: \"kubernetes.io/projected/30027bc7-2637-4abb-9568-1190cf80a3a5-kube-api-access-bhdgs\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.751975 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:49 crc kubenswrapper[4676]: I0124 00:21:49.751990 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30027bc7-2637-4abb-9568-1190cf80a3a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.110880 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wg6tc" event={"ID":"30027bc7-2637-4abb-9568-1190cf80a3a5","Type":"ContainerDied","Data":"3b055c8876511446c351694e38bb4288f579441b122767bc434100cd495602c9"} Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.111190 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b055c8876511446c351694e38bb4288f579441b122767bc434100cd495602c9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.110997 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wg6tc" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509471 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-z42h9"] Jan 24 00:21:50 crc kubenswrapper[4676]: E0124 00:21:50.509817 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a07672-60fb-4935-83e2-99f39e15427f" containerName="mariadb-account-create-update" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509829 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a07672-60fb-4935-83e2-99f39e15427f" containerName="mariadb-account-create-update" Jan 24 00:21:50 crc kubenswrapper[4676]: E0124 00:21:50.509843 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" containerName="init" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509849 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" containerName="init" Jan 24 00:21:50 crc kubenswrapper[4676]: E0124 00:21:50.509863 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" containerName="dnsmasq-dns" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509869 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" containerName="dnsmasq-dns" Jan 24 00:21:50 crc kubenswrapper[4676]: E0124 00:21:50.509876 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a61f15c0-2fa1-4d69-bb80-471d6e4e7e09" containerName="mariadb-account-create-update" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509882 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a61f15c0-2fa1-4d69-bb80-471d6e4e7e09" containerName="mariadb-account-create-update" Jan 24 00:21:50 crc kubenswrapper[4676]: E0124 00:21:50.509899 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4fb3ef-0cde-499c-8018-14cc96b495f5" containerName="mariadb-database-create" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509904 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4fb3ef-0cde-499c-8018-14cc96b495f5" containerName="mariadb-database-create" Jan 24 00:21:50 crc kubenswrapper[4676]: E0124 00:21:50.509921 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64efcf33-9ffe-402d-b0b0-9cf53c7a5495" containerName="mariadb-database-create" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509926 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="64efcf33-9ffe-402d-b0b0-9cf53c7a5495" containerName="mariadb-database-create" Jan 24 00:21:50 crc kubenswrapper[4676]: E0124 00:21:50.509934 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ace7871-3944-4c72-980f-c9d5e7d65c71" containerName="mariadb-account-create-update" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509940 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ace7871-3944-4c72-980f-c9d5e7d65c71" containerName="mariadb-account-create-update" Jan 24 00:21:50 crc kubenswrapper[4676]: E0124 00:21:50.509949 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30027bc7-2637-4abb-9568-1190cf80a3a5" containerName="keystone-db-sync" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509969 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="30027bc7-2637-4abb-9568-1190cf80a3a5" containerName="keystone-db-sync" Jan 24 00:21:50 crc kubenswrapper[4676]: E0124 00:21:50.509983 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b488c3f8-bc04-4c8d-98b7-72145ae9e948" containerName="mariadb-database-create" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.509990 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="b488c3f8-bc04-4c8d-98b7-72145ae9e948" containerName="mariadb-database-create" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.510122 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ace7871-3944-4c72-980f-c9d5e7d65c71" containerName="mariadb-account-create-update" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.510135 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="b488c3f8-bc04-4c8d-98b7-72145ae9e948" containerName="mariadb-database-create" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.510147 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="30027bc7-2637-4abb-9568-1190cf80a3a5" containerName="keystone-db-sync" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.510156 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="a61f15c0-2fa1-4d69-bb80-471d6e4e7e09" containerName="mariadb-account-create-update" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.510166 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="1905ca79-a4c4-4286-8d88-2855e7b9ba4c" containerName="dnsmasq-dns" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.510176 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="64a07672-60fb-4935-83e2-99f39e15427f" containerName="mariadb-account-create-update" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.510187 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4fb3ef-0cde-499c-8018-14cc96b495f5" containerName="mariadb-database-create" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.510198 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="64efcf33-9ffe-402d-b0b0-9cf53c7a5495" containerName="mariadb-database-create" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.510976 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.516273 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dbwnn"] Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.517185 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.525213 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.525559 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.525574 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.525741 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.536173 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fcgg9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.549273 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-z42h9"] Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.561448 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dbwnn"] Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.568986 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-config-data\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.569031 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-credential-keys\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.569073 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qhg5\" (UniqueName: \"kubernetes.io/projected/158d259b-be19-41c8-af3a-a84a3efeb214-kube-api-access-8qhg5\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.569167 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-fernet-keys\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.569857 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-scripts\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.569921 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-combined-ca-bundle\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671005 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-config-data\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671054 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-credential-keys\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671085 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671110 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qhg5\" (UniqueName: \"kubernetes.io/projected/158d259b-be19-41c8-af3a-a84a3efeb214-kube-api-access-8qhg5\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671136 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-config\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671153 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-fernet-keys\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671185 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-scripts\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671200 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671225 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-combined-ca-bundle\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671248 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671282 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brrz6\" (UniqueName: \"kubernetes.io/projected/2ac46337-c6f1-4c10-8a36-b9d36017920a-kube-api-access-brrz6\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.671300 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.679759 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-config-data\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.680552 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-scripts\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.683707 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-credential-keys\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.684339 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-combined-ca-bundle\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.684911 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-fernet-keys\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.723924 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qhg5\" (UniqueName: \"kubernetes.io/projected/158d259b-be19-41c8-af3a-a84a3efeb214-kube-api-access-8qhg5\") pod \"keystone-bootstrap-dbwnn\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.754620 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5ff79c7c45-k8zk9"] Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.768033 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.772365 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-config\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.772446 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.772484 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.772523 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brrz6\" (UniqueName: \"kubernetes.io/projected/2ac46337-c6f1-4c10-8a36-b9d36017920a-kube-api-access-brrz6\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.772542 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.772585 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.773619 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.773687 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.774216 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.774640 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-config\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.774800 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.782637 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.782911 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.798995 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-fmwfq" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.812040 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5ff79c7c45-k8zk9"] Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.828059 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.837239 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brrz6\" (UniqueName: \"kubernetes.io/projected/2ac46337-c6f1-4c10-8a36-b9d36017920a-kube-api-access-brrz6\") pod \"dnsmasq-dns-bbf5cc879-z42h9\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.840753 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.878824 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-scripts\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.878874 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-horizon-secret-key\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.878908 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-config-data\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.878944 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk5m5\" (UniqueName: \"kubernetes.io/projected/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-kube-api-access-xk5m5\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.878977 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-logs\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.895587 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-bpbdz"] Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.896689 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.922219 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xf4w5" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.922404 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.922515 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.923513 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bpbdz"] Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.987306 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-scripts\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.987350 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-horizon-secret-key\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.987576 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-config-data\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.987617 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk5m5\" (UniqueName: \"kubernetes.io/projected/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-kube-api-access-xk5m5\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.987652 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-logs\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.991408 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-logs\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.993792 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-config-data\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:50 crc kubenswrapper[4676]: I0124 00:21:50.996535 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-horizon-secret-key\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:50.998553 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-scripts\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.046440 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7klnv"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.047403 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.050016 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7klnv"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.070193 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-z42h9"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.070765 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.072364 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.072645 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-vxll9" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.072743 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.084043 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk5m5\" (UniqueName: \"kubernetes.io/projected/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-kube-api-access-xk5m5\") pod \"horizon-5ff79c7c45-k8zk9\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.085742 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-g5dnq"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.086597 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089511 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qvjc\" (UniqueName: \"kubernetes.io/projected/04db8ba1-c0de-4985-8a48-ead625786472-kube-api-access-7qvjc\") pod \"barbican-db-sync-g5dnq\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089556 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-combined-ca-bundle\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089604 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-combined-ca-bundle\") pod \"neutron-db-sync-bpbdz\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089626 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-etc-machine-id\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089661 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-scripts\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089693 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-db-sync-config-data\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089715 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-combined-ca-bundle\") pod \"barbican-db-sync-g5dnq\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089742 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-config-data\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089764 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-config\") pod \"neutron-db-sync-bpbdz\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089810 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mc9c\" (UniqueName: \"kubernetes.io/projected/01c23d24-6fd8-4660-a0f7-815fdf508f5b-kube-api-access-4mc9c\") pod \"neutron-db-sync-bpbdz\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089848 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-db-sync-config-data\") pod \"barbican-db-sync-g5dnq\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.089876 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsgpg\" (UniqueName: \"kubernetes.io/projected/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-kube-api-access-rsgpg\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.096850 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.132021 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-96w42" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.132325 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.165006 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-g5dnq"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195450 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-scripts\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195539 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-db-sync-config-data\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195575 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-combined-ca-bundle\") pod \"barbican-db-sync-g5dnq\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195608 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-config-data\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195641 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-config\") pod \"neutron-db-sync-bpbdz\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195694 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mc9c\" (UniqueName: \"kubernetes.io/projected/01c23d24-6fd8-4660-a0f7-815fdf508f5b-kube-api-access-4mc9c\") pod \"neutron-db-sync-bpbdz\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195746 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-db-sync-config-data\") pod \"barbican-db-sync-g5dnq\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195789 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsgpg\" (UniqueName: \"kubernetes.io/projected/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-kube-api-access-rsgpg\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195847 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qvjc\" (UniqueName: \"kubernetes.io/projected/04db8ba1-c0de-4985-8a48-ead625786472-kube-api-access-7qvjc\") pod \"barbican-db-sync-g5dnq\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195877 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-combined-ca-bundle\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.195927 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-combined-ca-bundle\") pod \"neutron-db-sync-bpbdz\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.196069 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-etc-machine-id\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.196267 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-etc-machine-id\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.225588 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-config-data\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.252524 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-db-sync-config-data\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.253840 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-config\") pod \"neutron-db-sync-bpbdz\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.254407 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-db-sync-config-data\") pod \"barbican-db-sync-g5dnq\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.254770 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-scripts\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.254894 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-combined-ca-bundle\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.305048 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-combined-ca-bundle\") pod \"barbican-db-sync-g5dnq\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.307486 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-combined-ca-bundle\") pod \"neutron-db-sync-bpbdz\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.322586 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-hpwcp"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.325108 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.345866 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsgpg\" (UniqueName: \"kubernetes.io/projected/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-kube-api-access-rsgpg\") pod \"cinder-db-sync-7klnv\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.377616 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7klnv" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.380195 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qvjc\" (UniqueName: \"kubernetes.io/projected/04db8ba1-c0de-4985-8a48-ead625786472-kube-api-access-7qvjc\") pod \"barbican-db-sync-g5dnq\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.389861 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.393749 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.421899 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.422079 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.424779 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-hpwcp"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.431231 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.431290 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.431429 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rjk5\" (UniqueName: \"kubernetes.io/projected/2417692e-7845-446e-a60a-bb3842851912-kube-api-access-5rjk5\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.431449 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.431484 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-config\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.431504 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.459764 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.460091 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.479074 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-4kdvp"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.480335 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.495439 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6qnnp" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.495826 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.495996 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.496310 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-694b9fc4d7-xkkn8"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.498052 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.515730 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.517078 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.519418 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-s5zsb" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.519625 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 24 00:21:51 crc kubenswrapper[4676]: I0124 00:21:51.519762 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.523808 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.530282 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4kdvp"] Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536179 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c17fc5e6-983e-4678-b22c-c68686271163-logs\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536243 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-config-data\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536263 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-config-data\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536294 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536336 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rjk5\" (UniqueName: \"kubernetes.io/projected/2417692e-7845-446e-a60a-bb3842851912-kube-api-access-5rjk5\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536355 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tljs7\" (UniqueName: \"kubernetes.io/projected/54d56910-d4b7-45b1-8699-5af7eaa29b96-kube-api-access-tljs7\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536371 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536409 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-scripts\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536423 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nv7wz\" (UniqueName: \"kubernetes.io/projected/c17fc5e6-983e-4678-b22c-c68686271163-kube-api-access-nv7wz\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536439 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-log-httpd\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536453 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-run-httpd\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536467 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-scripts\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536486 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-config\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536507 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536528 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-combined-ca-bundle\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536549 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536580 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.536606 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.537345 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.538081 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.538585 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-config\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.542158 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.550898 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.560583 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-694b9fc4d7-xkkn8"] Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.570925 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mc9c\" (UniqueName: \"kubernetes.io/projected/01c23d24-6fd8-4660-a0f7-815fdf508f5b-kube-api-access-4mc9c\") pod \"neutron-db-sync-bpbdz\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.571236 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.614239 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rjk5\" (UniqueName: \"kubernetes.io/projected/2417692e-7845-446e-a60a-bb3842851912-kube-api-access-5rjk5\") pod \"dnsmasq-dns-56df8fb6b7-hpwcp\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644417 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644459 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644502 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-logs\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644519 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c17fc5e6-983e-4678-b22c-c68686271163-logs\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644537 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644560 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-config-data\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644577 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd7b43d5-aa70-4c36-b1bb-4504bd262439-logs\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644610 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-scripts\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644636 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-config-data\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644653 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-config-data\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644673 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-487qg\" (UniqueName: \"kubernetes.io/projected/dd7b43d5-aa70-4c36-b1bb-4504bd262439-kube-api-access-487qg\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644696 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644717 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644735 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644760 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-config-data\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644776 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tljs7\" (UniqueName: \"kubernetes.io/projected/54d56910-d4b7-45b1-8699-5af7eaa29b96-kube-api-access-tljs7\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644794 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xjb8\" (UniqueName: \"kubernetes.io/projected/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-kube-api-access-8xjb8\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644814 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-scripts\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644832 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-scripts\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644847 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nv7wz\" (UniqueName: \"kubernetes.io/projected/c17fc5e6-983e-4678-b22c-c68686271163-kube-api-access-nv7wz\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644861 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-log-httpd\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644875 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-scripts\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644889 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-run-httpd\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644914 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd7b43d5-aa70-4c36-b1bb-4504bd262439-horizon-secret-key\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644919 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c17fc5e6-983e-4678-b22c-c68686271163-logs\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.644939 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-combined-ca-bundle\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.650114 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-log-httpd\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.657390 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-scripts\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.661955 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-run-httpd\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.695336 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.695813 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tljs7\" (UniqueName: \"kubernetes.io/projected/54d56910-d4b7-45b1-8699-5af7eaa29b96-kube-api-access-tljs7\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.707082 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-config-data\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.707929 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-scripts\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.707958 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.708327 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-combined-ca-bundle\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.721435 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.753991 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nv7wz\" (UniqueName: \"kubernetes.io/projected/c17fc5e6-983e-4678-b22c-c68686271163-kube-api-access-nv7wz\") pod \"placement-db-sync-4kdvp\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.760794 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd7b43d5-aa70-4c36-b1bb-4504bd262439-horizon-secret-key\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761269 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761313 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-logs\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761334 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761356 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-config-data\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761370 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd7b43d5-aa70-4c36-b1bb-4504bd262439-logs\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761402 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-scripts\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761446 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-487qg\" (UniqueName: \"kubernetes.io/projected/dd7b43d5-aa70-4c36-b1bb-4504bd262439-kube-api-access-487qg\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761468 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761491 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761514 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-config-data\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761532 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xjb8\" (UniqueName: \"kubernetes.io/projected/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-kube-api-access-8xjb8\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.761551 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-scripts\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.779542 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-logs\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.779750 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.796006 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd7b43d5-aa70-4c36-b1bb-4504bd262439-logs\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.796774 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.801138 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-config-data\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.801537 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-scripts\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.817894 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-config-data\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.818179 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd7b43d5-aa70-4c36-b1bb-4504bd262439-horizon-secret-key\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.819238 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-config-data\") pod \"ceilometer-0\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.827423 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.866922 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-487qg\" (UniqueName: \"kubernetes.io/projected/dd7b43d5-aa70-4c36-b1bb-4504bd262439-kube-api-access-487qg\") pod \"horizon-694b9fc4d7-xkkn8\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.898880 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.898943 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-scripts\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.899299 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.902399 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dbwnn"] Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.925553 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xjb8\" (UniqueName: \"kubernetes.io/projected/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-kube-api-access-8xjb8\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.928095 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4kdvp" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.962259 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:51.966891 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.013969 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.015624 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.031746 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.031917 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.059004 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.073959 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.092611 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.092786 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-662sq\" (UniqueName: \"kubernetes.io/projected/f5867251-cbde-4d93-8900-b134f99aae39-kube-api-access-662sq\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.092970 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.093020 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.093050 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.093190 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.093223 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.093334 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.194708 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.195252 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.195318 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-662sq\" (UniqueName: \"kubernetes.io/projected/f5867251-cbde-4d93-8900-b134f99aae39-kube-api-access-662sq\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.195433 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.195482 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.195505 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.195566 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.195590 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.199808 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dbwnn" event={"ID":"158d259b-be19-41c8-af3a-a84a3efeb214","Type":"ContainerStarted","Data":"7de28fe104be408633c60844298971c85a3dc1df0dbdf8b08da673fac8aac81a"} Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.199985 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.200205 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.200707 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-logs\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.209061 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.217144 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.218539 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.220235 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.225476 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-662sq\" (UniqueName: \"kubernetes.io/projected/f5867251-cbde-4d93-8900-b134f99aae39-kube-api-access-662sq\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.252726 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.283980 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:21:52 crc kubenswrapper[4676]: I0124 00:21:52.378036 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.127858 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5ff79c7c45-k8zk9"] Jan 24 00:21:53 crc kubenswrapper[4676]: W0124 00:21:53.134222 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ff71896_166b_4b5d_b11e_ebfb7d340e3d.slice/crio-dbda9f3c357ed36f88cb2516c4a0699d62520eb292d651052ab02beb2599dbcc WatchSource:0}: Error finding container dbda9f3c357ed36f88cb2516c4a0699d62520eb292d651052ab02beb2599dbcc: Status 404 returned error can't find the container with id dbda9f3c357ed36f88cb2516c4a0699d62520eb292d651052ab02beb2599dbcc Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.146979 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-z42h9"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.227835 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dbwnn" event={"ID":"158d259b-be19-41c8-af3a-a84a3efeb214","Type":"ContainerStarted","Data":"51869c6b7eeadd27a59a7b34889ba0093a949ad78c0444eec3606959ac5ba339"} Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.270720 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" event={"ID":"2ac46337-c6f1-4c10-8a36-b9d36017920a","Type":"ContainerStarted","Data":"69a7245a7f2fdb37f96d67c0c770f9ecb4530ed0f3864f0e13366799bfec1097"} Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.275213 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ff79c7c45-k8zk9" event={"ID":"4ff71896-166b-4b5d-b11e-ebfb7d340e3d","Type":"ContainerStarted","Data":"dbda9f3c357ed36f88cb2516c4a0699d62520eb292d651052ab02beb2599dbcc"} Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.275403 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.286218 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dbwnn" podStartSLOduration=3.286200661 podStartE2EDuration="3.286200661s" podCreationTimestamp="2026-01-24 00:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:53.283613362 +0000 UTC m=+1097.313584363" watchObservedRunningTime="2026-01-24 00:21:53.286200661 +0000 UTC m=+1097.316171652" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.354848 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5ff79c7c45-k8zk9"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.414162 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.437483 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-797567cdb9-7p8fk"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.438885 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.447731 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-797567cdb9-7p8fk"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.535431 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.538227 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-logs\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.538329 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-horizon-secret-key\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.538427 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-config-data\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.538484 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-scripts\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.538515 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn27t\" (UniqueName: \"kubernetes.io/projected/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-kube-api-access-hn27t\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.543646 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-694b9fc4d7-xkkn8"] Jan 24 00:21:53 crc kubenswrapper[4676]: W0124 00:21:53.559480 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd7b43d5_aa70_4c36_b1bb_4504bd262439.slice/crio-4152f9721423cdfc38209333d080f3609332e5c5cc25d9ded0d737267f382359 WatchSource:0}: Error finding container 4152f9721423cdfc38209333d080f3609332e5c5cc25d9ded0d737267f382359: Status 404 returned error can't find the container with id 4152f9721423cdfc38209333d080f3609332e5c5cc25d9ded0d737267f382359 Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.593643 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-bpbdz"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.616063 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7klnv"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.621588 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-hpwcp"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.639965 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-config-data\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.640018 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-scripts\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.646536 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn27t\" (UniqueName: \"kubernetes.io/projected/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-kube-api-access-hn27t\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.646615 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-logs\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.646833 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-horizon-secret-key\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.647893 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-logs\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.656429 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-scripts\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.658978 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-config-data\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.670794 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.695536 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-horizon-secret-key\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.729978 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn27t\" (UniqueName: \"kubernetes.io/projected/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-kube-api-access-hn27t\") pod \"horizon-797567cdb9-7p8fk\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.772697 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-4kdvp"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.779696 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.802193 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-g5dnq"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.842949 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:21:53 crc kubenswrapper[4676]: I0124 00:21:53.882424 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.297298 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-694b9fc4d7-xkkn8" event={"ID":"dd7b43d5-aa70-4c36-b1bb-4504bd262439","Type":"ContainerStarted","Data":"4152f9721423cdfc38209333d080f3609332e5c5cc25d9ded0d737267f382359"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.300436 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7klnv" event={"ID":"d1163cc6-7ce1-4f2b-9d4a-f3e215177842","Type":"ContainerStarted","Data":"70e37739b2a173b291b1d289a2d2839fa4f44eb620a1fa26acee32cd0b9f1716"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.302006 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fe9c2a69-3ecd-49b3-87d1-5c42a90af428","Type":"ContainerStarted","Data":"1824159d235b75a9dce59175487613a8beccbee50938b73c47eb45eda971ba07"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.303977 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bpbdz" event={"ID":"01c23d24-6fd8-4660-a0f7-815fdf508f5b","Type":"ContainerStarted","Data":"192824bb8ad730af6c36357704c067406da5b951731ba1841c964e1a78d10bc2"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.303999 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bpbdz" event={"ID":"01c23d24-6fd8-4660-a0f7-815fdf508f5b","Type":"ContainerStarted","Data":"21e76890e9eced941fb1d30013143862a35607ef8c3138b99e6cd707709e85c8"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.306037 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d56910-d4b7-45b1-8699-5af7eaa29b96","Type":"ContainerStarted","Data":"3a171779e2acfaaedcae6bf53bf013bc9cc89a659479afefe0a8349e50f12f73"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.310496 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4kdvp" event={"ID":"c17fc5e6-983e-4678-b22c-c68686271163","Type":"ContainerStarted","Data":"cb82773a377e4565776a6be3bc254bf675dcbd0497f37c26a350eb27b7526931"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.311691 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5867251-cbde-4d93-8900-b134f99aae39","Type":"ContainerStarted","Data":"8f9b82184da00b28ad31edc75c3a11bd122ec52ae047c31171199a425be304fc"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.313071 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-g5dnq" event={"ID":"04db8ba1-c0de-4985-8a48-ead625786472","Type":"ContainerStarted","Data":"534f710da5fb3c7907cdb2b5714b0d256a8306b5476b74f728dc51abaa10f9c0"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.318946 4676 generic.go:334] "Generic (PLEG): container finished" podID="2ac46337-c6f1-4c10-8a36-b9d36017920a" containerID="159d27826fe1136c6a706b48745620d8405fe60e7db1572dc165254bd5cc4a58" exitCode=0 Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.319216 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" event={"ID":"2ac46337-c6f1-4c10-8a36-b9d36017920a","Type":"ContainerDied","Data":"159d27826fe1136c6a706b48745620d8405fe60e7db1572dc165254bd5cc4a58"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.321129 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-bpbdz" podStartSLOduration=4.321119793 podStartE2EDuration="4.321119793s" podCreationTimestamp="2026-01-24 00:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:54.314704717 +0000 UTC m=+1098.344675718" watchObservedRunningTime="2026-01-24 00:21:54.321119793 +0000 UTC m=+1098.351090794" Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.328073 4676 generic.go:334] "Generic (PLEG): container finished" podID="2417692e-7845-446e-a60a-bb3842851912" containerID="4c0f072637105b041cb28fbed44761b77e903f1e8315f62df0bde2e50a63a7fa" exitCode=0 Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.329186 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" event={"ID":"2417692e-7845-446e-a60a-bb3842851912","Type":"ContainerDied","Data":"4c0f072637105b041cb28fbed44761b77e903f1e8315f62df0bde2e50a63a7fa"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.329217 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" event={"ID":"2417692e-7845-446e-a60a-bb3842851912","Type":"ContainerStarted","Data":"a17194b7f35eb9f8070fa9acebc74f7ada011686f763d785550450b94c66597c"} Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.424112 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-797567cdb9-7p8fk"] Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.845220 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.983984 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-svc\") pod \"2ac46337-c6f1-4c10-8a36-b9d36017920a\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.984362 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-swift-storage-0\") pod \"2ac46337-c6f1-4c10-8a36-b9d36017920a\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.984609 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brrz6\" (UniqueName: \"kubernetes.io/projected/2ac46337-c6f1-4c10-8a36-b9d36017920a-kube-api-access-brrz6\") pod \"2ac46337-c6f1-4c10-8a36-b9d36017920a\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.984659 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-sb\") pod \"2ac46337-c6f1-4c10-8a36-b9d36017920a\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.984685 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-config\") pod \"2ac46337-c6f1-4c10-8a36-b9d36017920a\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " Jan 24 00:21:54 crc kubenswrapper[4676]: I0124 00:21:54.984718 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-nb\") pod \"2ac46337-c6f1-4c10-8a36-b9d36017920a\" (UID: \"2ac46337-c6f1-4c10-8a36-b9d36017920a\") " Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.026111 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-config" (OuterVolumeSpecName: "config") pod "2ac46337-c6f1-4c10-8a36-b9d36017920a" (UID: "2ac46337-c6f1-4c10-8a36-b9d36017920a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.028440 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2ac46337-c6f1-4c10-8a36-b9d36017920a" (UID: "2ac46337-c6f1-4c10-8a36-b9d36017920a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.033734 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ac46337-c6f1-4c10-8a36-b9d36017920a-kube-api-access-brrz6" (OuterVolumeSpecName: "kube-api-access-brrz6") pod "2ac46337-c6f1-4c10-8a36-b9d36017920a" (UID: "2ac46337-c6f1-4c10-8a36-b9d36017920a"). InnerVolumeSpecName "kube-api-access-brrz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.050214 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2ac46337-c6f1-4c10-8a36-b9d36017920a" (UID: "2ac46337-c6f1-4c10-8a36-b9d36017920a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.073027 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2ac46337-c6f1-4c10-8a36-b9d36017920a" (UID: "2ac46337-c6f1-4c10-8a36-b9d36017920a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.088435 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brrz6\" (UniqueName: \"kubernetes.io/projected/2ac46337-c6f1-4c10-8a36-b9d36017920a-kube-api-access-brrz6\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.088473 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.088483 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.088492 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.088500 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.101271 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2ac46337-c6f1-4c10-8a36-b9d36017920a" (UID: "2ac46337-c6f1-4c10-8a36-b9d36017920a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.189802 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2ac46337-c6f1-4c10-8a36-b9d36017920a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.344920 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" event={"ID":"2ac46337-c6f1-4c10-8a36-b9d36017920a","Type":"ContainerDied","Data":"69a7245a7f2fdb37f96d67c0c770f9ecb4530ed0f3864f0e13366799bfec1097"} Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.345187 4676 scope.go:117] "RemoveContainer" containerID="159d27826fe1136c6a706b48745620d8405fe60e7db1572dc165254bd5cc4a58" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.345280 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-z42h9" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.350338 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fe9c2a69-3ecd-49b3-87d1-5c42a90af428","Type":"ContainerStarted","Data":"9b6d1f8ce083f680faa19785ed5bbaf58ec2053f4ead9760d1b30372e7ebb323"} Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.357190 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797567cdb9-7p8fk" event={"ID":"0b297a5e-7f42-435d-8c8e-fa8d113f8d74","Type":"ContainerStarted","Data":"d4a299117ca872f2981d137eaebb140c612fdabbc4e3e470ef6e031c80a0bce6"} Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.366594 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" event={"ID":"2417692e-7845-446e-a60a-bb3842851912","Type":"ContainerStarted","Data":"d6cc348312351f0aa6a3a89dece34eea57cb75b4f69f18c04d668c9984cdefc4"} Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.367183 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.389583 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5867251-cbde-4d93-8900-b134f99aae39","Type":"ContainerStarted","Data":"3f5282d9d816b83ab02ae74e2aab26c447e6e2a07343b6a6dc5c5f59c5e32711"} Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.432686 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-z42h9"] Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.444786 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-z42h9"] Jan 24 00:21:55 crc kubenswrapper[4676]: I0124 00:21:55.446559 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" podStartSLOduration=4.446542696 podStartE2EDuration="4.446542696s" podCreationTimestamp="2026-01-24 00:21:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:55.420280882 +0000 UTC m=+1099.450251873" watchObservedRunningTime="2026-01-24 00:21:55.446542696 +0000 UTC m=+1099.476513697" Jan 24 00:21:56 crc kubenswrapper[4676]: I0124 00:21:56.293256 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ac46337-c6f1-4c10-8a36-b9d36017920a" path="/var/lib/kubelet/pods/2ac46337-c6f1-4c10-8a36-b9d36017920a/volumes" Jan 24 00:21:57 crc kubenswrapper[4676]: I0124 00:21:57.432947 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fe9c2a69-3ecd-49b3-87d1-5c42a90af428","Type":"ContainerStarted","Data":"3b215804cc138ac23930129381e72626b89a03c7f85d05eb606c6576d28a097c"} Jan 24 00:21:57 crc kubenswrapper[4676]: I0124 00:21:57.433078 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerName="glance-log" containerID="cri-o://9b6d1f8ce083f680faa19785ed5bbaf58ec2053f4ead9760d1b30372e7ebb323" gracePeriod=30 Jan 24 00:21:57 crc kubenswrapper[4676]: I0124 00:21:57.433343 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerName="glance-httpd" containerID="cri-o://3b215804cc138ac23930129381e72626b89a03c7f85d05eb606c6576d28a097c" gracePeriod=30 Jan 24 00:21:57 crc kubenswrapper[4676]: I0124 00:21:57.435840 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5867251-cbde-4d93-8900-b134f99aae39","Type":"ContainerStarted","Data":"bc729c98cba3454809e024f8daa025b45fade86eac51870c7e12abcef362e249"} Jan 24 00:21:57 crc kubenswrapper[4676]: I0124 00:21:57.435983 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f5867251-cbde-4d93-8900-b134f99aae39" containerName="glance-log" containerID="cri-o://3f5282d9d816b83ab02ae74e2aab26c447e6e2a07343b6a6dc5c5f59c5e32711" gracePeriod=30 Jan 24 00:21:57 crc kubenswrapper[4676]: I0124 00:21:57.435995 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f5867251-cbde-4d93-8900-b134f99aae39" containerName="glance-httpd" containerID="cri-o://bc729c98cba3454809e024f8daa025b45fade86eac51870c7e12abcef362e249" gracePeriod=30 Jan 24 00:21:57 crc kubenswrapper[4676]: I0124 00:21:57.453869 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.453853955 podStartE2EDuration="6.453853955s" podCreationTimestamp="2026-01-24 00:21:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:57.451443711 +0000 UTC m=+1101.481414712" watchObservedRunningTime="2026-01-24 00:21:57.453853955 +0000 UTC m=+1101.483824956" Jan 24 00:21:58 crc kubenswrapper[4676]: I0124 00:21:58.466887 4676 generic.go:334] "Generic (PLEG): container finished" podID="f5867251-cbde-4d93-8900-b134f99aae39" containerID="bc729c98cba3454809e024f8daa025b45fade86eac51870c7e12abcef362e249" exitCode=0 Jan 24 00:21:58 crc kubenswrapper[4676]: I0124 00:21:58.467191 4676 generic.go:334] "Generic (PLEG): container finished" podID="f5867251-cbde-4d93-8900-b134f99aae39" containerID="3f5282d9d816b83ab02ae74e2aab26c447e6e2a07343b6a6dc5c5f59c5e32711" exitCode=143 Jan 24 00:21:58 crc kubenswrapper[4676]: I0124 00:21:58.467083 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5867251-cbde-4d93-8900-b134f99aae39","Type":"ContainerDied","Data":"bc729c98cba3454809e024f8daa025b45fade86eac51870c7e12abcef362e249"} Jan 24 00:21:58 crc kubenswrapper[4676]: I0124 00:21:58.467271 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5867251-cbde-4d93-8900-b134f99aae39","Type":"ContainerDied","Data":"3f5282d9d816b83ab02ae74e2aab26c447e6e2a07343b6a6dc5c5f59c5e32711"} Jan 24 00:21:58 crc kubenswrapper[4676]: I0124 00:21:58.477067 4676 generic.go:334] "Generic (PLEG): container finished" podID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerID="3b215804cc138ac23930129381e72626b89a03c7f85d05eb606c6576d28a097c" exitCode=0 Jan 24 00:21:58 crc kubenswrapper[4676]: I0124 00:21:58.477106 4676 generic.go:334] "Generic (PLEG): container finished" podID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerID="9b6d1f8ce083f680faa19785ed5bbaf58ec2053f4ead9760d1b30372e7ebb323" exitCode=143 Jan 24 00:21:58 crc kubenswrapper[4676]: I0124 00:21:58.477132 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fe9c2a69-3ecd-49b3-87d1-5c42a90af428","Type":"ContainerDied","Data":"3b215804cc138ac23930129381e72626b89a03c7f85d05eb606c6576d28a097c"} Jan 24 00:21:58 crc kubenswrapper[4676]: I0124 00:21:58.477170 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fe9c2a69-3ecd-49b3-87d1-5c42a90af428","Type":"ContainerDied","Data":"9b6d1f8ce083f680faa19785ed5bbaf58ec2053f4ead9760d1b30372e7ebb323"} Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.487149 4676 generic.go:334] "Generic (PLEG): container finished" podID="158d259b-be19-41c8-af3a-a84a3efeb214" containerID="51869c6b7eeadd27a59a7b34889ba0093a949ad78c0444eec3606959ac5ba339" exitCode=0 Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.487470 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dbwnn" event={"ID":"158d259b-be19-41c8-af3a-a84a3efeb214","Type":"ContainerDied","Data":"51869c6b7eeadd27a59a7b34889ba0093a949ad78c0444eec3606959ac5ba339"} Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.507591 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=9.507573425 podStartE2EDuration="9.507573425s" podCreationTimestamp="2026-01-24 00:21:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:21:57.487760323 +0000 UTC m=+1101.517731324" watchObservedRunningTime="2026-01-24 00:21:59.507573425 +0000 UTC m=+1103.537544426" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.777564 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-694b9fc4d7-xkkn8"] Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.809764 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-bf988b4bd-ls7hp"] Jan 24 00:21:59 crc kubenswrapper[4676]: E0124 00:21:59.813322 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ac46337-c6f1-4c10-8a36-b9d36017920a" containerName="init" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.814459 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ac46337-c6f1-4c10-8a36-b9d36017920a" containerName="init" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.815361 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ac46337-c6f1-4c10-8a36-b9d36017920a" containerName="init" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.816302 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.824292 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.851934 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bf988b4bd-ls7hp"] Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.949348 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-797567cdb9-7p8fk"] Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.988434 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f876ddf46-fs7qv"] Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.990301 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.993774 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwffk\" (UniqueName: \"kubernetes.io/projected/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-kube-api-access-cwffk\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.993848 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-secret-key\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.993877 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-config-data\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.993914 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-tls-certs\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.994439 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-logs\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.994505 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-combined-ca-bundle\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:21:59 crc kubenswrapper[4676]: I0124 00:21:59.994569 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-scripts\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.006187 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f876ddf46-fs7qv"] Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.095794 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-horizon-secret-key\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.095835 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-secret-key\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.095861 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-config-data\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.095889 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-tls-certs\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.095905 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-logs\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.095925 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99tbt\" (UniqueName: \"kubernetes.io/projected/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-kube-api-access-99tbt\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.095975 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-logs\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.095996 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-combined-ca-bundle\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.096011 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-scripts\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.096031 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-horizon-tls-certs\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.096052 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-config-data\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.096067 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-scripts\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.096093 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-combined-ca-bundle\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.096114 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwffk\" (UniqueName: \"kubernetes.io/projected/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-kube-api-access-cwffk\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.101526 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-combined-ca-bundle\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.102694 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-config-data\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.105716 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-tls-certs\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.105843 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-logs\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.106021 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-scripts\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.111768 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwffk\" (UniqueName: \"kubernetes.io/projected/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-kube-api-access-cwffk\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.115961 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-secret-key\") pod \"horizon-bf988b4bd-ls7hp\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.171420 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.198036 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-config-data\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.198097 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-combined-ca-bundle\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.198167 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-horizon-secret-key\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.198210 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-logs\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.198253 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99tbt\" (UniqueName: \"kubernetes.io/projected/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-kube-api-access-99tbt\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.198338 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-scripts\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.198358 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-horizon-tls-certs\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.199716 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-config-data\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.202225 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-combined-ca-bundle\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.202251 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-logs\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.207831 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-horizon-secret-key\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.212250 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-scripts\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.213511 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-horizon-tls-certs\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.213851 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99tbt\" (UniqueName: \"kubernetes.io/projected/ac7dce6b-3bd9-4ad9-9485-83d9384b8bad-kube-api-access-99tbt\") pod \"horizon-f876ddf46-fs7qv\" (UID: \"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad\") " pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:00 crc kubenswrapper[4676]: I0124 00:22:00.320986 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:01 crc kubenswrapper[4676]: I0124 00:22:01.697924 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:22:01 crc kubenswrapper[4676]: I0124 00:22:01.815146 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-p2f7j"] Jan 24 00:22:01 crc kubenswrapper[4676]: I0124 00:22:01.815404 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="dnsmasq-dns" containerID="cri-o://eecf36d183fa31e03de5954a2aec194cfc8588694d23fe41c8598d42bf4437a3" gracePeriod=10 Jan 24 00:22:03 crc kubenswrapper[4676]: I0124 00:22:03.522743 4676 generic.go:334] "Generic (PLEG): container finished" podID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerID="eecf36d183fa31e03de5954a2aec194cfc8588694d23fe41c8598d42bf4437a3" exitCode=0 Jan 24 00:22:03 crc kubenswrapper[4676]: I0124 00:22:03.522819 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" event={"ID":"92c7f906-f96b-4d55-b771-6d6425b32b85","Type":"ContainerDied","Data":"eecf36d183fa31e03de5954a2aec194cfc8588694d23fe41c8598d42bf4437a3"} Jan 24 00:22:04 crc kubenswrapper[4676]: I0124 00:22:04.543335 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: connect: connection refused" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.777461 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.780972 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.782338 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.851568 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-combined-ca-bundle\") pod \"f5867251-cbde-4d93-8900-b134f99aae39\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.851974 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-combined-ca-bundle\") pod \"158d259b-be19-41c8-af3a-a84a3efeb214\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852010 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-logs\") pod \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852097 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852126 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-httpd-run\") pod \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852157 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-combined-ca-bundle\") pod \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852185 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-logs\") pod \"f5867251-cbde-4d93-8900-b134f99aae39\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852223 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-public-tls-certs\") pod \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852257 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-scripts\") pod \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852282 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-config-data\") pod \"f5867251-cbde-4d93-8900-b134f99aae39\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852312 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-scripts\") pod \"f5867251-cbde-4d93-8900-b134f99aae39\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852356 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-scripts\") pod \"158d259b-be19-41c8-af3a-a84a3efeb214\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852425 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xjb8\" (UniqueName: \"kubernetes.io/projected/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-kube-api-access-8xjb8\") pod \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852456 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-fernet-keys\") pod \"158d259b-be19-41c8-af3a-a84a3efeb214\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852484 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-config-data\") pod \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\" (UID: \"fe9c2a69-3ecd-49b3-87d1-5c42a90af428\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852517 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qhg5\" (UniqueName: \"kubernetes.io/projected/158d259b-be19-41c8-af3a-a84a3efeb214-kube-api-access-8qhg5\") pod \"158d259b-be19-41c8-af3a-a84a3efeb214\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852561 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-internal-tls-certs\") pod \"f5867251-cbde-4d93-8900-b134f99aae39\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852584 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-662sq\" (UniqueName: \"kubernetes.io/projected/f5867251-cbde-4d93-8900-b134f99aae39-kube-api-access-662sq\") pod \"f5867251-cbde-4d93-8900-b134f99aae39\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852615 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-credential-keys\") pod \"158d259b-be19-41c8-af3a-a84a3efeb214\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852638 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-config-data\") pod \"158d259b-be19-41c8-af3a-a84a3efeb214\" (UID: \"158d259b-be19-41c8-af3a-a84a3efeb214\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852657 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"f5867251-cbde-4d93-8900-b134f99aae39\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.852679 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-httpd-run\") pod \"f5867251-cbde-4d93-8900-b134f99aae39\" (UID: \"f5867251-cbde-4d93-8900-b134f99aae39\") " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.854080 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f5867251-cbde-4d93-8900-b134f99aae39" (UID: "f5867251-cbde-4d93-8900-b134f99aae39"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.874722 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "158d259b-be19-41c8-af3a-a84a3efeb214" (UID: "158d259b-be19-41c8-af3a-a84a3efeb214"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.874894 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5867251-cbde-4d93-8900-b134f99aae39-kube-api-access-662sq" (OuterVolumeSpecName: "kube-api-access-662sq") pod "f5867251-cbde-4d93-8900-b134f99aae39" (UID: "f5867251-cbde-4d93-8900-b134f99aae39"). InnerVolumeSpecName "kube-api-access-662sq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.876817 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "158d259b-be19-41c8-af3a-a84a3efeb214" (UID: "158d259b-be19-41c8-af3a-a84a3efeb214"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.880155 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-logs" (OuterVolumeSpecName: "logs") pod "f5867251-cbde-4d93-8900-b134f99aae39" (UID: "f5867251-cbde-4d93-8900-b134f99aae39"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.880446 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-logs" (OuterVolumeSpecName: "logs") pod "fe9c2a69-3ecd-49b3-87d1-5c42a90af428" (UID: "fe9c2a69-3ecd-49b3-87d1-5c42a90af428"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.881609 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fe9c2a69-3ecd-49b3-87d1-5c42a90af428" (UID: "fe9c2a69-3ecd-49b3-87d1-5c42a90af428"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.888732 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "fe9c2a69-3ecd-49b3-87d1-5c42a90af428" (UID: "fe9c2a69-3ecd-49b3-87d1-5c42a90af428"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.888822 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "f5867251-cbde-4d93-8900-b134f99aae39" (UID: "f5867251-cbde-4d93-8900-b134f99aae39"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.889512 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-scripts" (OuterVolumeSpecName: "scripts") pod "158d259b-be19-41c8-af3a-a84a3efeb214" (UID: "158d259b-be19-41c8-af3a-a84a3efeb214"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.893319 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-kube-api-access-8xjb8" (OuterVolumeSpecName: "kube-api-access-8xjb8") pod "fe9c2a69-3ecd-49b3-87d1-5c42a90af428" (UID: "fe9c2a69-3ecd-49b3-87d1-5c42a90af428"). InnerVolumeSpecName "kube-api-access-8xjb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.893443 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/158d259b-be19-41c8-af3a-a84a3efeb214-kube-api-access-8qhg5" (OuterVolumeSpecName: "kube-api-access-8qhg5") pod "158d259b-be19-41c8-af3a-a84a3efeb214" (UID: "158d259b-be19-41c8-af3a-a84a3efeb214"). InnerVolumeSpecName "kube-api-access-8qhg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.898531 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-scripts" (OuterVolumeSpecName: "scripts") pod "fe9c2a69-3ecd-49b3-87d1-5c42a90af428" (UID: "fe9c2a69-3ecd-49b3-87d1-5c42a90af428"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.911559 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-scripts" (OuterVolumeSpecName: "scripts") pod "f5867251-cbde-4d93-8900-b134f99aae39" (UID: "f5867251-cbde-4d93-8900-b134f99aae39"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.941987 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "158d259b-be19-41c8-af3a-a84a3efeb214" (UID: "158d259b-be19-41c8-af3a-a84a3efeb214"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.946908 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5867251-cbde-4d93-8900-b134f99aae39" (UID: "f5867251-cbde-4d93-8900-b134f99aae39"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.954951 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.954977 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.954987 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955008 4676 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955017 4676 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955102 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955114 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955122 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955129 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955138 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xjb8\" (UniqueName: \"kubernetes.io/projected/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-kube-api-access-8xjb8\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955148 4676 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955157 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qhg5\" (UniqueName: \"kubernetes.io/projected/158d259b-be19-41c8-af3a-a84a3efeb214-kube-api-access-8qhg5\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955177 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-662sq\" (UniqueName: \"kubernetes.io/projected/f5867251-cbde-4d93-8900-b134f99aae39-kube-api-access-662sq\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955185 4676 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955199 4676 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.955208 4676 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f5867251-cbde-4d93-8900-b134f99aae39-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.956507 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe9c2a69-3ecd-49b3-87d1-5c42a90af428" (UID: "fe9c2a69-3ecd-49b3-87d1-5c42a90af428"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.959038 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-config-data" (OuterVolumeSpecName: "config-data") pod "158d259b-be19-41c8-af3a-a84a3efeb214" (UID: "158d259b-be19-41c8-af3a-a84a3efeb214"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.982256 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-config-data" (OuterVolumeSpecName: "config-data") pod "f5867251-cbde-4d93-8900-b134f99aae39" (UID: "f5867251-cbde-4d93-8900-b134f99aae39"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.983696 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f5867251-cbde-4d93-8900-b134f99aae39" (UID: "f5867251-cbde-4d93-8900-b134f99aae39"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.993526 4676 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.994717 4676 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 24 00:22:05 crc kubenswrapper[4676]: I0124 00:22:05.997666 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-config-data" (OuterVolumeSpecName: "config-data") pod "fe9c2a69-3ecd-49b3-87d1-5c42a90af428" (UID: "fe9c2a69-3ecd-49b3-87d1-5c42a90af428"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.005743 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fe9c2a69-3ecd-49b3-87d1-5c42a90af428" (UID: "fe9c2a69-3ecd-49b3-87d1-5c42a90af428"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.057069 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.057110 4676 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.057131 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158d259b-be19-41c8-af3a-a84a3efeb214-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.057144 4676 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.057156 4676 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.057167 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.057181 4676 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9c2a69-3ecd-49b3-87d1-5c42a90af428-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.057193 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5867251-cbde-4d93-8900-b134f99aae39-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.569187 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f5867251-cbde-4d93-8900-b134f99aae39","Type":"ContainerDied","Data":"8f9b82184da00b28ad31edc75c3a11bd122ec52ae047c31171199a425be304fc"} Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.569235 4676 scope.go:117] "RemoveContainer" containerID="bc729c98cba3454809e024f8daa025b45fade86eac51870c7e12abcef362e249" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.569337 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.576300 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dbwnn" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.576464 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dbwnn" event={"ID":"158d259b-be19-41c8-af3a-a84a3efeb214","Type":"ContainerDied","Data":"7de28fe104be408633c60844298971c85a3dc1df0dbdf8b08da673fac8aac81a"} Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.576499 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7de28fe104be408633c60844298971c85a3dc1df0dbdf8b08da673fac8aac81a" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.581443 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fe9c2a69-3ecd-49b3-87d1-5c42a90af428","Type":"ContainerDied","Data":"1824159d235b75a9dce59175487613a8beccbee50938b73c47eb45eda971ba07"} Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.581489 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.605486 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.680137 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.693497 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.709409 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.720881 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:22:06 crc kubenswrapper[4676]: E0124 00:22:06.721314 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerName="glance-log" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.721338 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerName="glance-log" Jan 24 00:22:06 crc kubenswrapper[4676]: E0124 00:22:06.721361 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="158d259b-be19-41c8-af3a-a84a3efeb214" containerName="keystone-bootstrap" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.721370 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="158d259b-be19-41c8-af3a-a84a3efeb214" containerName="keystone-bootstrap" Jan 24 00:22:06 crc kubenswrapper[4676]: E0124 00:22:06.721402 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerName="glance-httpd" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.721411 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerName="glance-httpd" Jan 24 00:22:06 crc kubenswrapper[4676]: E0124 00:22:06.721427 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5867251-cbde-4d93-8900-b134f99aae39" containerName="glance-httpd" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.721435 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5867251-cbde-4d93-8900-b134f99aae39" containerName="glance-httpd" Jan 24 00:22:06 crc kubenswrapper[4676]: E0124 00:22:06.721445 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5867251-cbde-4d93-8900-b134f99aae39" containerName="glance-log" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.721453 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5867251-cbde-4d93-8900-b134f99aae39" containerName="glance-log" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.725012 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerName="glance-httpd" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.725047 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" containerName="glance-log" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.725067 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5867251-cbde-4d93-8900-b134f99aae39" containerName="glance-log" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.725076 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5867251-cbde-4d93-8900-b134f99aae39" containerName="glance-httpd" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.725088 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="158d259b-be19-41c8-af3a-a84a3efeb214" containerName="keystone-bootstrap" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.726218 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.728288 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.729005 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.729832 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.729968 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-s5zsb" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.738337 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.739733 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.741474 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.741921 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.758617 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.782218 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.872902 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.872946 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blvls\" (UniqueName: \"kubernetes.io/projected/643e6d41-6572-4f21-8651-7f577967bfe8-kube-api-access-blvls\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.872975 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.872998 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-config-data\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873014 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-scripts\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873035 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-logs\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873057 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-config-data\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873082 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873098 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-logs\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873122 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49fqw\" (UniqueName: \"kubernetes.io/projected/77ec0563-46f1-45b0-892b-352d088f9517-kube-api-access-49fqw\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873139 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-scripts\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873165 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873272 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873312 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873422 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.873587 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.943291 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dbwnn"] Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.950211 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dbwnn"] Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975090 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975142 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blvls\" (UniqueName: \"kubernetes.io/projected/643e6d41-6572-4f21-8651-7f577967bfe8-kube-api-access-blvls\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975174 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975195 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-config-data\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975210 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-scripts\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975229 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-logs\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975251 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-config-data\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975276 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975294 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-logs\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975317 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49fqw\" (UniqueName: \"kubernetes.io/projected/77ec0563-46f1-45b0-892b-352d088f9517-kube-api-access-49fqw\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975334 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-scripts\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975360 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975398 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975415 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975449 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.975473 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.981590 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.982012 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-logs\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.983242 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.983489 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.983619 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.983759 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.984071 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.985274 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-logs\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.989803 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-config-data\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.990720 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.994992 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-scripts\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.996554 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-scripts\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:06 crc kubenswrapper[4676]: I0124 00:22:06.997226 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-config-data\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.002671 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.002769 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49fqw\" (UniqueName: \"kubernetes.io/projected/77ec0563-46f1-45b0-892b-352d088f9517-kube-api-access-49fqw\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.010661 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blvls\" (UniqueName: \"kubernetes.io/projected/643e6d41-6572-4f21-8651-7f577967bfe8-kube-api-access-blvls\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.014555 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " pod="openstack/glance-default-external-api-0" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.033579 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.055827 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-l49lm"] Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.056877 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.059873 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.060118 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.060263 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.060495 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fcgg9" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.060683 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.065862 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.077829 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.078050 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-l49lm"] Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.178950 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-credential-keys\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.179001 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjz2c\" (UniqueName: \"kubernetes.io/projected/a7327fe7-179a-4492-8823-94b1067c17d4-kube-api-access-wjz2c\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.179023 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-scripts\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.179130 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-combined-ca-bundle\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.179183 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-config-data\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.179231 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-fernet-keys\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.286345 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-fernet-keys\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.286494 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-credential-keys\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.286524 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjz2c\" (UniqueName: \"kubernetes.io/projected/a7327fe7-179a-4492-8823-94b1067c17d4-kube-api-access-wjz2c\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.286546 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-scripts\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.286616 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-combined-ca-bundle\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.286654 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-config-data\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.290588 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-scripts\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.292082 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-config-data\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.302286 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-fernet-keys\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.302915 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-combined-ca-bundle\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.320266 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-credential-keys\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.339663 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjz2c\" (UniqueName: \"kubernetes.io/projected/a7327fe7-179a-4492-8823-94b1067c17d4-kube-api-access-wjz2c\") pod \"keystone-bootstrap-l49lm\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:07 crc kubenswrapper[4676]: I0124 00:22:07.434830 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:08 crc kubenswrapper[4676]: I0124 00:22:08.264348 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="158d259b-be19-41c8-af3a-a84a3efeb214" path="/var/lib/kubelet/pods/158d259b-be19-41c8-af3a-a84a3efeb214/volumes" Jan 24 00:22:08 crc kubenswrapper[4676]: I0124 00:22:08.265446 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5867251-cbde-4d93-8900-b134f99aae39" path="/var/lib/kubelet/pods/f5867251-cbde-4d93-8900-b134f99aae39/volumes" Jan 24 00:22:08 crc kubenswrapper[4676]: I0124 00:22:08.266482 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe9c2a69-3ecd-49b3-87d1-5c42a90af428" path="/var/lib/kubelet/pods/fe9c2a69-3ecd-49b3-87d1-5c42a90af428/volumes" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.095736 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.096657 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbdh8bh5cdh64fh9bh6ch5f5h7dh5cbh5f7h86h5b6h8bh5cch567h68dh596h5ch548h666h5ch569hddh57fh5d6hcch684hc9h667h56dh659hd4q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xk5m5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-5ff79c7c45-k8zk9_openstack(4ff71896-166b-4b5d-b11e-ebfb7d340e3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.098101 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.098296 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n67ch5fhb4h554h599h5f5h5d4h5d6h699hb4h5dh8dh5dfh79h74h5ch64ch666h668h665h5b9h68fh699h598h67bh655h569h55h584h665h649h87q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hn27t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-797567cdb9-7p8fk_openstack(0b297a5e-7f42-435d-8c8e-fa8d113f8d74): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.100481 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-5ff79c7c45-k8zk9" podUID="4ff71896-166b-4b5d-b11e-ebfb7d340e3d" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.100965 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-797567cdb9-7p8fk" podUID="0b297a5e-7f42-435d-8c8e-fa8d113f8d74" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.117100 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.117330 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d6h656h596hc8h5ddhbdh5d9h68h564h678h554h99hb5h68ch77h8h584h6bh5cbh679hddh5fbh64h5f6h58h669h5b8h5fbh688h59fh58fh9cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-487qg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-694b9fc4d7-xkkn8_openstack(dd7b43d5-aa70-4c36-b1bb-4504bd262439): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.119354 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-694b9fc4d7-xkkn8" podUID="dd7b43d5-aa70-4c36-b1bb-4504bd262439" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.567750 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.567907 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7qvjc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-g5dnq_openstack(04db8ba1-c0de-4985-8a48-ead625786472): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.569088 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-g5dnq" podUID="04db8ba1-c0de-4985-8a48-ead625786472" Jan 24 00:22:11 crc kubenswrapper[4676]: E0124 00:22:11.622678 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-g5dnq" podUID="04db8ba1-c0de-4985-8a48-ead625786472" Jan 24 00:22:14 crc kubenswrapper[4676]: I0124 00:22:14.543192 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: i/o timeout" Jan 24 00:22:19 crc kubenswrapper[4676]: I0124 00:22:19.544704 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: i/o timeout" Jan 24 00:22:19 crc kubenswrapper[4676]: I0124 00:22:19.545529 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.327391 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.334170 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.343711 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.357640 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487124 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-swift-storage-0\") pod \"92c7f906-f96b-4d55-b771-6d6425b32b85\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487188 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-logs\") pod \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487215 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-config\") pod \"92c7f906-f96b-4d55-b771-6d6425b32b85\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487243 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-sb\") pod \"92c7f906-f96b-4d55-b771-6d6425b32b85\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487263 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd7b43d5-aa70-4c36-b1bb-4504bd262439-logs\") pod \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487316 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-scripts\") pod \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487366 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-config-data\") pod \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487413 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-svc\") pod \"92c7f906-f96b-4d55-b771-6d6425b32b85\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487467 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-487qg\" (UniqueName: \"kubernetes.io/projected/dd7b43d5-aa70-4c36-b1bb-4504bd262439-kube-api-access-487qg\") pod \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487499 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-logs\") pod \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487534 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-config-data\") pod \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487560 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-scripts\") pod \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487609 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-horizon-secret-key\") pod \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487634 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd7b43d5-aa70-4c36-b1bb-4504bd262439-horizon-secret-key\") pod \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487665 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-horizon-secret-key\") pod \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487686 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-config-data\") pod \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487733 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-nb\") pod \"92c7f906-f96b-4d55-b771-6d6425b32b85\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487779 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk5m5\" (UniqueName: \"kubernetes.io/projected/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-kube-api-access-xk5m5\") pod \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\" (UID: \"4ff71896-166b-4b5d-b11e-ebfb7d340e3d\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487811 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-scripts\") pod \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\" (UID: \"dd7b43d5-aa70-4c36-b1bb-4504bd262439\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487839 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hn27t\" (UniqueName: \"kubernetes.io/projected/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-kube-api-access-hn27t\") pod \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\" (UID: \"0b297a5e-7f42-435d-8c8e-fa8d113f8d74\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.487876 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68mnw\" (UniqueName: \"kubernetes.io/projected/92c7f906-f96b-4d55-b771-6d6425b32b85-kube-api-access-68mnw\") pod \"92c7f906-f96b-4d55-b771-6d6425b32b85\" (UID: \"92c7f906-f96b-4d55-b771-6d6425b32b85\") " Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.488315 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-scripts" (OuterVolumeSpecName: "scripts") pod "0b297a5e-7f42-435d-8c8e-fa8d113f8d74" (UID: "0b297a5e-7f42-435d-8c8e-fa8d113f8d74"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.488504 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd7b43d5-aa70-4c36-b1bb-4504bd262439-logs" (OuterVolumeSpecName: "logs") pod "dd7b43d5-aa70-4c36-b1bb-4504bd262439" (UID: "dd7b43d5-aa70-4c36-b1bb-4504bd262439"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.489225 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-scripts" (OuterVolumeSpecName: "scripts") pod "dd7b43d5-aa70-4c36-b1bb-4504bd262439" (UID: "dd7b43d5-aa70-4c36-b1bb-4504bd262439"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.489604 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-config-data" (OuterVolumeSpecName: "config-data") pod "4ff71896-166b-4b5d-b11e-ebfb7d340e3d" (UID: "4ff71896-166b-4b5d-b11e-ebfb7d340e3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.489769 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-logs" (OuterVolumeSpecName: "logs") pod "4ff71896-166b-4b5d-b11e-ebfb7d340e3d" (UID: "4ff71896-166b-4b5d-b11e-ebfb7d340e3d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.489935 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-config-data" (OuterVolumeSpecName: "config-data") pod "dd7b43d5-aa70-4c36-b1bb-4504bd262439" (UID: "dd7b43d5-aa70-4c36-b1bb-4504bd262439"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.493296 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd7b43d5-aa70-4c36-b1bb-4504bd262439-kube-api-access-487qg" (OuterVolumeSpecName: "kube-api-access-487qg") pod "dd7b43d5-aa70-4c36-b1bb-4504bd262439" (UID: "dd7b43d5-aa70-4c36-b1bb-4504bd262439"). InnerVolumeSpecName "kube-api-access-487qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.493656 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-kube-api-access-xk5m5" (OuterVolumeSpecName: "kube-api-access-xk5m5") pod "4ff71896-166b-4b5d-b11e-ebfb7d340e3d" (UID: "4ff71896-166b-4b5d-b11e-ebfb7d340e3d"). InnerVolumeSpecName "kube-api-access-xk5m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.493681 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-scripts" (OuterVolumeSpecName: "scripts") pod "4ff71896-166b-4b5d-b11e-ebfb7d340e3d" (UID: "4ff71896-166b-4b5d-b11e-ebfb7d340e3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.494138 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-logs" (OuterVolumeSpecName: "logs") pod "0b297a5e-7f42-435d-8c8e-fa8d113f8d74" (UID: "0b297a5e-7f42-435d-8c8e-fa8d113f8d74"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.494607 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-config-data" (OuterVolumeSpecName: "config-data") pod "0b297a5e-7f42-435d-8c8e-fa8d113f8d74" (UID: "0b297a5e-7f42-435d-8c8e-fa8d113f8d74"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.495349 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-kube-api-access-hn27t" (OuterVolumeSpecName: "kube-api-access-hn27t") pod "0b297a5e-7f42-435d-8c8e-fa8d113f8d74" (UID: "0b297a5e-7f42-435d-8c8e-fa8d113f8d74"). InnerVolumeSpecName "kube-api-access-hn27t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.497107 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92c7f906-f96b-4d55-b771-6d6425b32b85-kube-api-access-68mnw" (OuterVolumeSpecName: "kube-api-access-68mnw") pod "92c7f906-f96b-4d55-b771-6d6425b32b85" (UID: "92c7f906-f96b-4d55-b771-6d6425b32b85"). InnerVolumeSpecName "kube-api-access-68mnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.508982 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0b297a5e-7f42-435d-8c8e-fa8d113f8d74" (UID: "0b297a5e-7f42-435d-8c8e-fa8d113f8d74"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.509702 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "4ff71896-166b-4b5d-b11e-ebfb7d340e3d" (UID: "4ff71896-166b-4b5d-b11e-ebfb7d340e3d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.511978 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd7b43d5-aa70-4c36-b1bb-4504bd262439-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "dd7b43d5-aa70-4c36-b1bb-4504bd262439" (UID: "dd7b43d5-aa70-4c36-b1bb-4504bd262439"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.528899 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "92c7f906-f96b-4d55-b771-6d6425b32b85" (UID: "92c7f906-f96b-4d55-b771-6d6425b32b85"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.532689 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-config" (OuterVolumeSpecName: "config") pod "92c7f906-f96b-4d55-b771-6d6425b32b85" (UID: "92c7f906-f96b-4d55-b771-6d6425b32b85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.536118 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "92c7f906-f96b-4d55-b771-6d6425b32b85" (UID: "92c7f906-f96b-4d55-b771-6d6425b32b85"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.542236 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "92c7f906-f96b-4d55-b771-6d6425b32b85" (UID: "92c7f906-f96b-4d55-b771-6d6425b32b85"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.561668 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "92c7f906-f96b-4d55-b771-6d6425b32b85" (UID: "92c7f906-f96b-4d55-b771-6d6425b32b85"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591202 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591246 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591261 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591271 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591281 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dd7b43d5-aa70-4c36-b1bb-4504bd262439-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591290 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591323 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591331 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591340 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-487qg\" (UniqueName: \"kubernetes.io/projected/dd7b43d5-aa70-4c36-b1bb-4504bd262439-kube-api-access-487qg\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591351 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591359 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591367 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591387 4676 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591394 4676 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dd7b43d5-aa70-4c36-b1bb-4504bd262439-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591403 4676 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591429 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591436 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c7f906-f96b-4d55-b771-6d6425b32b85-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591444 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk5m5\" (UniqueName: \"kubernetes.io/projected/4ff71896-166b-4b5d-b11e-ebfb7d340e3d-kube-api-access-xk5m5\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591451 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd7b43d5-aa70-4c36-b1bb-4504bd262439-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591460 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hn27t\" (UniqueName: \"kubernetes.io/projected/0b297a5e-7f42-435d-8c8e-fa8d113f8d74-kube-api-access-hn27t\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.591468 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68mnw\" (UniqueName: \"kubernetes.io/projected/92c7f906-f96b-4d55-b771-6d6425b32b85-kube-api-access-68mnw\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.730351 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-797567cdb9-7p8fk" event={"ID":"0b297a5e-7f42-435d-8c8e-fa8d113f8d74","Type":"ContainerDied","Data":"d4a299117ca872f2981d137eaebb140c612fdabbc4e3e470ef6e031c80a0bce6"} Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.730642 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-797567cdb9-7p8fk" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.743519 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-694b9fc4d7-xkkn8" event={"ID":"dd7b43d5-aa70-4c36-b1bb-4504bd262439","Type":"ContainerDied","Data":"4152f9721423cdfc38209333d080f3609332e5c5cc25d9ded0d737267f382359"} Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.743536 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-694b9fc4d7-xkkn8" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.751434 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5ff79c7c45-k8zk9" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.751450 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5ff79c7c45-k8zk9" event={"ID":"4ff71896-166b-4b5d-b11e-ebfb7d340e3d","Type":"ContainerDied","Data":"dbda9f3c357ed36f88cb2516c4a0699d62520eb292d651052ab02beb2599dbcc"} Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.754357 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" event={"ID":"92c7f906-f96b-4d55-b771-6d6425b32b85","Type":"ContainerDied","Data":"af5442d400f3ec8197d8bc02bc11e29bed515b4e2a1c69855707d8adbf0ab335"} Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.754436 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.810141 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-797567cdb9-7p8fk"] Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.830153 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-797567cdb9-7p8fk"] Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.849828 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-p2f7j"] Jan 24 00:22:22 crc kubenswrapper[4676]: E0124 00:22:22.852303 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Jan 24 00:22:22 crc kubenswrapper[4676]: E0124 00:22:22.852490 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65fh8hfbh5b5h596h58fh6ch66fh585h546h5c6h4h568h6fh5bdh68h5f6h658h75h64bh577h645h545h65ch677h684h587h77h649h694h64fh7q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tljs7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(54d56910-d4b7-45b1-8699-5af7eaa29b96): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.869916 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-p2f7j"] Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.915463 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-694b9fc4d7-xkkn8"] Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.927333 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-694b9fc4d7-xkkn8"] Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.940673 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5ff79c7c45-k8zk9"] Jan 24 00:22:22 crc kubenswrapper[4676]: I0124 00:22:22.946743 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5ff79c7c45-k8zk9"] Jan 24 00:22:23 crc kubenswrapper[4676]: I0124 00:22:23.767449 4676 generic.go:334] "Generic (PLEG): container finished" podID="01c23d24-6fd8-4660-a0f7-815fdf508f5b" containerID="192824bb8ad730af6c36357704c067406da5b951731ba1841c964e1a78d10bc2" exitCode=0 Jan 24 00:22:23 crc kubenswrapper[4676]: I0124 00:22:23.767541 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bpbdz" event={"ID":"01c23d24-6fd8-4660-a0f7-815fdf508f5b","Type":"ContainerDied","Data":"192824bb8ad730af6c36357704c067406da5b951731ba1841c964e1a78d10bc2"} Jan 24 00:22:23 crc kubenswrapper[4676]: I0124 00:22:23.822879 4676 scope.go:117] "RemoveContainer" containerID="3f5282d9d816b83ab02ae74e2aab26c447e6e2a07343b6a6dc5c5f59c5e32711" Jan 24 00:22:23 crc kubenswrapper[4676]: E0124 00:22:23.852187 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Jan 24 00:22:23 crc kubenswrapper[4676]: E0124 00:22:23.852583 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rsgpg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7klnv_openstack(d1163cc6-7ce1-4f2b-9d4a-f3e215177842): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:22:23 crc kubenswrapper[4676]: E0124 00:22:23.853819 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7klnv" podUID="d1163cc6-7ce1-4f2b-9d4a-f3e215177842" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.018511 4676 scope.go:117] "RemoveContainer" containerID="3b215804cc138ac23930129381e72626b89a03c7f85d05eb606c6576d28a097c" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.099340 4676 scope.go:117] "RemoveContainer" containerID="9b6d1f8ce083f680faa19785ed5bbaf58ec2053f4ead9760d1b30372e7ebb323" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.152491 4676 scope.go:117] "RemoveContainer" containerID="eecf36d183fa31e03de5954a2aec194cfc8588694d23fe41c8598d42bf4437a3" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.182189 4676 scope.go:117] "RemoveContainer" containerID="041801085cb8948d9dc29d8e4969e7d85a03615218568807e333c3bd95e75583" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.273152 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b297a5e-7f42-435d-8c8e-fa8d113f8d74" path="/var/lib/kubelet/pods/0b297a5e-7f42-435d-8c8e-fa8d113f8d74/volumes" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.273618 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ff71896-166b-4b5d-b11e-ebfb7d340e3d" path="/var/lib/kubelet/pods/4ff71896-166b-4b5d-b11e-ebfb7d340e3d/volumes" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.273995 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" path="/var/lib/kubelet/pods/92c7f906-f96b-4d55-b771-6d6425b32b85/volumes" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.274609 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd7b43d5-aa70-4c36-b1bb-4504bd262439" path="/var/lib/kubelet/pods/dd7b43d5-aa70-4c36-b1bb-4504bd262439/volumes" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.396292 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-l49lm"] Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.405339 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f876ddf46-fs7qv"] Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.412138 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-bf988b4bd-ls7hp"] Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.545560 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-p2f7j" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.127:5353: i/o timeout" Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.605776 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:22:24 crc kubenswrapper[4676]: W0124 00:22:24.676675 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ec0563_46f1_45b0_892b_352d088f9517.slice/crio-07d4384811e1f2313a45aceab532176f3c3d05348e79e5d4955b534eb2e2868c WatchSource:0}: Error finding container 07d4384811e1f2313a45aceab532176f3c3d05348e79e5d4955b534eb2e2868c: Status 404 returned error can't find the container with id 07d4384811e1f2313a45aceab532176f3c3d05348e79e5d4955b534eb2e2868c Jan 24 00:22:24 crc kubenswrapper[4676]: W0124 00:22:24.689580 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac7dce6b_3bd9_4ad9_9485_83d9384b8bad.slice/crio-9bb78a03e6699d79a2bf36213743a6fe1a49edf74f9acac38e1b0834e4817a68 WatchSource:0}: Error finding container 9bb78a03e6699d79a2bf36213743a6fe1a49edf74f9acac38e1b0834e4817a68: Status 404 returned error can't find the container with id 9bb78a03e6699d79a2bf36213743a6fe1a49edf74f9acac38e1b0834e4817a68 Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.786015 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l49lm" event={"ID":"a7327fe7-179a-4492-8823-94b1067c17d4","Type":"ContainerStarted","Data":"02e78984834a58ca6226d3263c50f11bea3fc65e0cf4f6d946dcccb6c1c331db"} Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.789836 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"77ec0563-46f1-45b0-892b-352d088f9517","Type":"ContainerStarted","Data":"07d4384811e1f2313a45aceab532176f3c3d05348e79e5d4955b534eb2e2868c"} Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.795971 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4kdvp" event={"ID":"c17fc5e6-983e-4678-b22c-c68686271163","Type":"ContainerStarted","Data":"531b26a23fe05a6b22c6e35ea5ff32eb5bb18c34495f3ce046a990e3e2b684b5"} Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.806850 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf988b4bd-ls7hp" event={"ID":"9d2451a0-4896-46e4-9b9e-e309ccdf02f2","Type":"ContainerStarted","Data":"e16c7aef4fd28d5aab8b64abddb94def56344a74657a3085b40c7035b723d59e"} Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.808351 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f876ddf46-fs7qv" event={"ID":"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad","Type":"ContainerStarted","Data":"9bb78a03e6699d79a2bf36213743a6fe1a49edf74f9acac38e1b0834e4817a68"} Jan 24 00:22:24 crc kubenswrapper[4676]: I0124 00:22:24.820799 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-4kdvp" podStartSLOduration=5.296070105 podStartE2EDuration="33.820783714s" podCreationTimestamp="2026-01-24 00:21:51 +0000 UTC" firstStartedPulling="2026-01-24 00:21:53.694398031 +0000 UTC m=+1097.724369032" lastFinishedPulling="2026-01-24 00:22:22.21911164 +0000 UTC m=+1126.249082641" observedRunningTime="2026-01-24 00:22:24.817631408 +0000 UTC m=+1128.847602409" watchObservedRunningTime="2026-01-24 00:22:24.820783714 +0000 UTC m=+1128.850754715" Jan 24 00:22:24 crc kubenswrapper[4676]: E0124 00:22:24.828838 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-7klnv" podUID="d1163cc6-7ce1-4f2b-9d4a-f3e215177842" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.208545 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.281532 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-combined-ca-bundle\") pod \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.281699 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-config\") pod \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.281721 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mc9c\" (UniqueName: \"kubernetes.io/projected/01c23d24-6fd8-4660-a0f7-815fdf508f5b-kube-api-access-4mc9c\") pod \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\" (UID: \"01c23d24-6fd8-4660-a0f7-815fdf508f5b\") " Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.295126 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01c23d24-6fd8-4660-a0f7-815fdf508f5b-kube-api-access-4mc9c" (OuterVolumeSpecName: "kube-api-access-4mc9c") pod "01c23d24-6fd8-4660-a0f7-815fdf508f5b" (UID: "01c23d24-6fd8-4660-a0f7-815fdf508f5b"). InnerVolumeSpecName "kube-api-access-4mc9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.296968 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.352359 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01c23d24-6fd8-4660-a0f7-815fdf508f5b" (UID: "01c23d24-6fd8-4660-a0f7-815fdf508f5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.369480 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-config" (OuterVolumeSpecName: "config") pod "01c23d24-6fd8-4660-a0f7-815fdf508f5b" (UID: "01c23d24-6fd8-4660-a0f7-815fdf508f5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.393324 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.393356 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mc9c\" (UniqueName: \"kubernetes.io/projected/01c23d24-6fd8-4660-a0f7-815fdf508f5b-kube-api-access-4mc9c\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.393400 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01c23d24-6fd8-4660-a0f7-815fdf508f5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.906717 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf988b4bd-ls7hp" event={"ID":"9d2451a0-4896-46e4-9b9e-e309ccdf02f2","Type":"ContainerStarted","Data":"ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb"} Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.923093 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f876ddf46-fs7qv" event={"ID":"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad","Type":"ContainerStarted","Data":"bcd4f4f533f53af3d30042718f9bbd0d4265fa269bd879cf2e257782f033723e"} Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.940601 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"643e6d41-6572-4f21-8651-7f577967bfe8","Type":"ContainerStarted","Data":"7eaab757b11ce1880975cf769167065cfd261eeb0ede6ebf82f2d90ebe083765"} Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.949821 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-bpbdz" event={"ID":"01c23d24-6fd8-4660-a0f7-815fdf508f5b","Type":"ContainerDied","Data":"21e76890e9eced941fb1d30013143862a35607ef8c3138b99e6cd707709e85c8"} Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.949860 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21e76890e9eced941fb1d30013143862a35607ef8c3138b99e6cd707709e85c8" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.949947 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-bpbdz" Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.991888 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d56910-d4b7-45b1-8699-5af7eaa29b96","Type":"ContainerStarted","Data":"f53a3c895796c44b94d61775eb726821f040873261a1a8fbdb79f578ca085ec5"} Jan 24 00:22:25 crc kubenswrapper[4676]: I0124 00:22:25.997947 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l49lm" event={"ID":"a7327fe7-179a-4492-8823-94b1067c17d4","Type":"ContainerStarted","Data":"47e15b8611d25a16d27aae65f29f85b602c09a46ed5cfeeea3914ac200a1d0a7"} Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.017082 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"77ec0563-46f1-45b0-892b-352d088f9517","Type":"ContainerStarted","Data":"7c7d07ff8323e9ccead03e4d66a066ad64d46892c577bd47375ca06cbd3b2418"} Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.061120 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-l49lm" podStartSLOduration=19.061098386 podStartE2EDuration="19.061098386s" podCreationTimestamp="2026-01-24 00:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:26.056924529 +0000 UTC m=+1130.086895590" watchObservedRunningTime="2026-01-24 00:22:26.061098386 +0000 UTC m=+1130.091069397" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.129876 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-r22sn"] Jan 24 00:22:26 crc kubenswrapper[4676]: E0124 00:22:26.130223 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="init" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.130239 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="init" Jan 24 00:22:26 crc kubenswrapper[4676]: E0124 00:22:26.130249 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="dnsmasq-dns" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.130256 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="dnsmasq-dns" Jan 24 00:22:26 crc kubenswrapper[4676]: E0124 00:22:26.130264 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01c23d24-6fd8-4660-a0f7-815fdf508f5b" containerName="neutron-db-sync" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.130270 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="01c23d24-6fd8-4660-a0f7-815fdf508f5b" containerName="neutron-db-sync" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.130456 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="92c7f906-f96b-4d55-b771-6d6425b32b85" containerName="dnsmasq-dns" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.130472 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="01c23d24-6fd8-4660-a0f7-815fdf508f5b" containerName="neutron-db-sync" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.131238 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.193929 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-r22sn"] Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.304287 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-756b7b5794-xhgs6"] Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.305537 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.311814 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.312064 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xf4w5" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.312267 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.312388 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.320239 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.320301 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-svc\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.320325 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.320368 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq96k\" (UniqueName: \"kubernetes.io/projected/da4decd9-74a7-4128-a34b-e30de6318a09-kube-api-access-cq96k\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.320415 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-config\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.320431 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.321075 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-756b7b5794-xhgs6"] Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426195 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-config\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426257 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426309 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkwcg\" (UniqueName: \"kubernetes.io/projected/1d557c78-075a-44f7-a530-860ae3ec8ffd-kube-api-access-kkwcg\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426364 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-config\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426399 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-combined-ca-bundle\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426468 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-httpd-config\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426489 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-ovndb-tls-certs\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426507 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426554 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-svc\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426577 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.426627 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq96k\" (UniqueName: \"kubernetes.io/projected/da4decd9-74a7-4128-a34b-e30de6318a09-kube-api-access-cq96k\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.427078 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-config\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.452666 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq96k\" (UniqueName: \"kubernetes.io/projected/da4decd9-74a7-4128-a34b-e30de6318a09-kube-api-access-cq96k\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.467768 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.468528 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.468685 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-svc\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.469145 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-r22sn\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.527594 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkwcg\" (UniqueName: \"kubernetes.io/projected/1d557c78-075a-44f7-a530-860ae3ec8ffd-kube-api-access-kkwcg\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.527644 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-config\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.527661 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-combined-ca-bundle\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.527677 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-httpd-config\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.527689 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-ovndb-tls-certs\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.532636 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-config\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.544323 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-httpd-config\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.544841 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-ovndb-tls-certs\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.545254 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-combined-ca-bundle\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.550329 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkwcg\" (UniqueName: \"kubernetes.io/projected/1d557c78-075a-44f7-a530-860ae3ec8ffd-kube-api-access-kkwcg\") pod \"neutron-756b7b5794-xhgs6\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.661888 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:26 crc kubenswrapper[4676]: I0124 00:22:26.758949 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.046677 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"77ec0563-46f1-45b0-892b-352d088f9517","Type":"ContainerStarted","Data":"1989fb85f0013ac25e4d8fb0c72cdef0a1a08075037f574a25a6701244b8dd50"} Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.053245 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf988b4bd-ls7hp" event={"ID":"9d2451a0-4896-46e4-9b9e-e309ccdf02f2","Type":"ContainerStarted","Data":"9b6a2140f368edfea407714ce9a9afb7bd7057997de5cfab1861b7a16523894e"} Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.067748 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.067804 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.110186 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=21.11017168 podStartE2EDuration="21.11017168s" podCreationTimestamp="2026-01-24 00:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:27.080124007 +0000 UTC m=+1131.110095008" watchObservedRunningTime="2026-01-24 00:22:27.11017168 +0000 UTC m=+1131.140142681" Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.111899 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f876ddf46-fs7qv" event={"ID":"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad","Type":"ContainerStarted","Data":"c8c367e4ea3e593fa82e51f67b7f8371ac3844af82364461f69e4513b638a905"} Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.113206 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-bf988b4bd-ls7hp" podStartSLOduration=27.480296154 podStartE2EDuration="28.113200942s" podCreationTimestamp="2026-01-24 00:21:59 +0000 UTC" firstStartedPulling="2026-01-24 00:22:24.706441117 +0000 UTC m=+1128.736412108" lastFinishedPulling="2026-01-24 00:22:25.339345895 +0000 UTC m=+1129.369316896" observedRunningTime="2026-01-24 00:22:27.110325275 +0000 UTC m=+1131.140296276" watchObservedRunningTime="2026-01-24 00:22:27.113200942 +0000 UTC m=+1131.143171943" Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.138786 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f876ddf46-fs7qv" podStartSLOduration=27.513597696 podStartE2EDuration="28.1387688s" podCreationTimestamp="2026-01-24 00:21:59 +0000 UTC" firstStartedPulling="2026-01-24 00:22:24.706322963 +0000 UTC m=+1128.736293964" lastFinishedPulling="2026-01-24 00:22:25.331494067 +0000 UTC m=+1129.361465068" observedRunningTime="2026-01-24 00:22:27.135415658 +0000 UTC m=+1131.165386659" watchObservedRunningTime="2026-01-24 00:22:27.1387688 +0000 UTC m=+1131.168739801" Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.188841 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"643e6d41-6572-4f21-8651-7f577967bfe8","Type":"ContainerStarted","Data":"b6416b75b9719da0d296633782ac2fc04e1dd20b2e9d142f5025be8c5a5d754d"} Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.209208 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.232086 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.447906 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-756b7b5794-xhgs6"] Jan 24 00:22:27 crc kubenswrapper[4676]: I0124 00:22:27.801055 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-r22sn"] Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.197146 4676 generic.go:334] "Generic (PLEG): container finished" podID="c17fc5e6-983e-4678-b22c-c68686271163" containerID="531b26a23fe05a6b22c6e35ea5ff32eb5bb18c34495f3ce046a990e3e2b684b5" exitCode=0 Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.197347 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4kdvp" event={"ID":"c17fc5e6-983e-4678-b22c-c68686271163","Type":"ContainerDied","Data":"531b26a23fe05a6b22c6e35ea5ff32eb5bb18c34495f3ce046a990e3e2b684b5"} Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.201913 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" event={"ID":"da4decd9-74a7-4128-a34b-e30de6318a09","Type":"ContainerStarted","Data":"a839f941eaa0eb3fef79a831fc52f7ec100168032ef563b034304acff5b0a94c"} Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.215960 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"643e6d41-6572-4f21-8651-7f577967bfe8","Type":"ContainerStarted","Data":"e3769b8edac359e6c5b4b6e492f9b2bb7ee70731b0b4c2778ff42c7ddf474ff8"} Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.220190 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756b7b5794-xhgs6" event={"ID":"1d557c78-075a-44f7-a530-860ae3ec8ffd","Type":"ContainerStarted","Data":"ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a"} Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.220221 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.220233 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756b7b5794-xhgs6" event={"ID":"1d557c78-075a-44f7-a530-860ae3ec8ffd","Type":"ContainerStarted","Data":"90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e"} Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.220246 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.220254 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756b7b5794-xhgs6" event={"ID":"1d557c78-075a-44f7-a530-860ae3ec8ffd","Type":"ContainerStarted","Data":"2a37de0875eb36fe5f67aef1544f7cf8a8b0cd17d7552f5a5cd3065c861671b8"} Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.221040 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.260638 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=22.260618258 podStartE2EDuration="22.260618258s" podCreationTimestamp="2026-01-24 00:22:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:28.233759132 +0000 UTC m=+1132.263730133" watchObservedRunningTime="2026-01-24 00:22:28.260618258 +0000 UTC m=+1132.290589259" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.270788 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-756b7b5794-xhgs6" podStartSLOduration=2.270774938 podStartE2EDuration="2.270774938s" podCreationTimestamp="2026-01-24 00:22:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:28.264142166 +0000 UTC m=+1132.294113167" watchObservedRunningTime="2026-01-24 00:22:28.270774938 +0000 UTC m=+1132.300745939" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.621932 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b7b989687-7nkv4"] Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.625407 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.631393 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.632053 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.654424 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b7b989687-7nkv4"] Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.693867 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-public-tls-certs\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.693924 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-httpd-config\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.693982 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-config\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.694037 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-ovndb-tls-certs\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.694060 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-internal-tls-certs\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.694083 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-combined-ca-bundle\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.694266 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm9l4\" (UniqueName: \"kubernetes.io/projected/2e3ec968-d892-487a-930a-44c79123d54b-kube-api-access-nm9l4\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.796422 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-config\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.796504 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-ovndb-tls-certs\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.796528 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-internal-tls-certs\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.796554 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-combined-ca-bundle\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.796592 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm9l4\" (UniqueName: \"kubernetes.io/projected/2e3ec968-d892-487a-930a-44c79123d54b-kube-api-access-nm9l4\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.796648 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-public-tls-certs\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.796675 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-httpd-config\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.801300 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-combined-ca-bundle\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.805055 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-public-tls-certs\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.806342 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-ovndb-tls-certs\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.806395 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-config\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.808301 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-internal-tls-certs\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.809159 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-httpd-config\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.812063 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm9l4\" (UniqueName: \"kubernetes.io/projected/2e3ec968-d892-487a-930a-44c79123d54b-kube-api-access-nm9l4\") pod \"neutron-b7b989687-7nkv4\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:28 crc kubenswrapper[4676]: I0124 00:22:28.960812 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.259909 4676 generic.go:334] "Generic (PLEG): container finished" podID="da4decd9-74a7-4128-a34b-e30de6318a09" containerID="8f47f3dbe46f5da8dd23954963ad8dc0cd2663f72dd53c26d9fe4db7d0341294" exitCode=0 Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.259972 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" event={"ID":"da4decd9-74a7-4128-a34b-e30de6318a09","Type":"ContainerDied","Data":"8f47f3dbe46f5da8dd23954963ad8dc0cd2663f72dd53c26d9fe4db7d0341294"} Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.268405 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-g5dnq" event={"ID":"04db8ba1-c0de-4985-8a48-ead625786472","Type":"ContainerStarted","Data":"9f879edfb9e19e2ba5d1f7acc73cb74d671c4cd334c91721768ab3acc3bba0e9"} Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.289519 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-g5dnq" podStartSLOduration=4.217343798 podStartE2EDuration="38.289482669s" podCreationTimestamp="2026-01-24 00:21:51 +0000 UTC" firstStartedPulling="2026-01-24 00:21:53.693594727 +0000 UTC m=+1097.723565728" lastFinishedPulling="2026-01-24 00:22:27.765733608 +0000 UTC m=+1131.795704599" observedRunningTime="2026-01-24 00:22:29.284107376 +0000 UTC m=+1133.314078377" watchObservedRunningTime="2026-01-24 00:22:29.289482669 +0000 UTC m=+1133.319453670" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.591044 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4kdvp" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.709832 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-scripts\") pod \"c17fc5e6-983e-4678-b22c-c68686271163\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.710008 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-combined-ca-bundle\") pod \"c17fc5e6-983e-4678-b22c-c68686271163\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.710068 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c17fc5e6-983e-4678-b22c-c68686271163-logs\") pod \"c17fc5e6-983e-4678-b22c-c68686271163\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.710086 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-config-data\") pod \"c17fc5e6-983e-4678-b22c-c68686271163\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.710157 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nv7wz\" (UniqueName: \"kubernetes.io/projected/c17fc5e6-983e-4678-b22c-c68686271163-kube-api-access-nv7wz\") pod \"c17fc5e6-983e-4678-b22c-c68686271163\" (UID: \"c17fc5e6-983e-4678-b22c-c68686271163\") " Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.710430 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c17fc5e6-983e-4678-b22c-c68686271163-logs" (OuterVolumeSpecName: "logs") pod "c17fc5e6-983e-4678-b22c-c68686271163" (UID: "c17fc5e6-983e-4678-b22c-c68686271163"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.711028 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c17fc5e6-983e-4678-b22c-c68686271163-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.715238 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-scripts" (OuterVolumeSpecName: "scripts") pod "c17fc5e6-983e-4678-b22c-c68686271163" (UID: "c17fc5e6-983e-4678-b22c-c68686271163"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.719203 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c17fc5e6-983e-4678-b22c-c68686271163-kube-api-access-nv7wz" (OuterVolumeSpecName: "kube-api-access-nv7wz") pod "c17fc5e6-983e-4678-b22c-c68686271163" (UID: "c17fc5e6-983e-4678-b22c-c68686271163"). InnerVolumeSpecName "kube-api-access-nv7wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.740677 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c17fc5e6-983e-4678-b22c-c68686271163" (UID: "c17fc5e6-983e-4678-b22c-c68686271163"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.752692 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-config-data" (OuterVolumeSpecName: "config-data") pod "c17fc5e6-983e-4678-b22c-c68686271163" (UID: "c17fc5e6-983e-4678-b22c-c68686271163"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.812732 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.812770 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.812783 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c17fc5e6-983e-4678-b22c-c68686271163-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.812792 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nv7wz\" (UniqueName: \"kubernetes.io/projected/c17fc5e6-983e-4678-b22c-c68686271163-kube-api-access-nv7wz\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:29 crc kubenswrapper[4676]: I0124 00:22:29.873629 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b7b989687-7nkv4"] Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.172024 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.172274 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.332099 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b7b989687-7nkv4" event={"ID":"2e3ec968-d892-487a-930a-44c79123d54b","Type":"ContainerStarted","Data":"59d763231e986383c1ba6301030139cf376841912045d87d7f5c53b19c30ff2a"} Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.332139 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.332155 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.335862 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-4kdvp" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.337385 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-4kdvp" event={"ID":"c17fc5e6-983e-4678-b22c-c68686271163","Type":"ContainerDied","Data":"cb82773a377e4565776a6be3bc254bf675dcbd0497f37c26a350eb27b7526931"} Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.337428 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb82773a377e4565776a6be3bc254bf675dcbd0497f37c26a350eb27b7526931" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.528050 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-765f6cdf58-5q9v9"] Jan 24 00:22:30 crc kubenswrapper[4676]: E0124 00:22:30.528466 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c17fc5e6-983e-4678-b22c-c68686271163" containerName="placement-db-sync" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.528485 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="c17fc5e6-983e-4678-b22c-c68686271163" containerName="placement-db-sync" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.528636 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="c17fc5e6-983e-4678-b22c-c68686271163" containerName="placement-db-sync" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.529593 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.535071 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.537117 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.537327 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.537488 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6qnnp" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.537578 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.560643 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-765f6cdf58-5q9v9"] Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.630276 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-internal-tls-certs\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.630654 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-public-tls-certs\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.630680 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-combined-ca-bundle\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.630733 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-scripts\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.630769 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e63383-223a-4247-8589-03ab5a33f980-logs\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.630821 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ftlq\" (UniqueName: \"kubernetes.io/projected/21e63383-223a-4247-8589-03ab5a33f980-kube-api-access-2ftlq\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.630845 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-config-data\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.732124 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e63383-223a-4247-8589-03ab5a33f980-logs\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.732208 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ftlq\" (UniqueName: \"kubernetes.io/projected/21e63383-223a-4247-8589-03ab5a33f980-kube-api-access-2ftlq\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.732235 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-config-data\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.732271 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-internal-tls-certs\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.732299 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-public-tls-certs\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.732318 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-combined-ca-bundle\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.732365 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-scripts\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.738907 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-scripts\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.739165 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-config-data\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.739223 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21e63383-223a-4247-8589-03ab5a33f980-logs\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.740039 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-internal-tls-certs\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.742191 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-combined-ca-bundle\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.742814 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21e63383-223a-4247-8589-03ab5a33f980-public-tls-certs\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.764056 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ftlq\" (UniqueName: \"kubernetes.io/projected/21e63383-223a-4247-8589-03ab5a33f980-kube-api-access-2ftlq\") pod \"placement-765f6cdf58-5q9v9\" (UID: \"21e63383-223a-4247-8589-03ab5a33f980\") " pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:30 crc kubenswrapper[4676]: I0124 00:22:30.866487 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:31 crc kubenswrapper[4676]: I0124 00:22:31.374554 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b7b989687-7nkv4" event={"ID":"2e3ec968-d892-487a-930a-44c79123d54b","Type":"ContainerStarted","Data":"b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae"} Jan 24 00:22:31 crc kubenswrapper[4676]: I0124 00:22:31.374852 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b7b989687-7nkv4" event={"ID":"2e3ec968-d892-487a-930a-44c79123d54b","Type":"ContainerStarted","Data":"bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724"} Jan 24 00:22:31 crc kubenswrapper[4676]: I0124 00:22:31.375232 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:22:31 crc kubenswrapper[4676]: I0124 00:22:31.402937 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-b7b989687-7nkv4" podStartSLOduration=3.402922424 podStartE2EDuration="3.402922424s" podCreationTimestamp="2026-01-24 00:22:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:31.401112709 +0000 UTC m=+1135.431083710" watchObservedRunningTime="2026-01-24 00:22:31.402922424 +0000 UTC m=+1135.432893425" Jan 24 00:22:31 crc kubenswrapper[4676]: I0124 00:22:31.616208 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-765f6cdf58-5q9v9"] Jan 24 00:22:33 crc kubenswrapper[4676]: I0124 00:22:33.404748 4676 generic.go:334] "Generic (PLEG): container finished" podID="a7327fe7-179a-4492-8823-94b1067c17d4" containerID="47e15b8611d25a16d27aae65f29f85b602c09a46ed5cfeeea3914ac200a1d0a7" exitCode=0 Jan 24 00:22:33 crc kubenswrapper[4676]: I0124 00:22:33.406769 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l49lm" event={"ID":"a7327fe7-179a-4492-8823-94b1067c17d4","Type":"ContainerDied","Data":"47e15b8611d25a16d27aae65f29f85b602c09a46ed5cfeeea3914ac200a1d0a7"} Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.415408 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d56910-d4b7-45b1-8699-5af7eaa29b96","Type":"ContainerStarted","Data":"dae61f7a8060ccdb486fd53e86c1fca33bbb71e0fb9f0d456ad2e0a5a8007cbd"} Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.417457 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" event={"ID":"da4decd9-74a7-4128-a34b-e30de6318a09","Type":"ContainerStarted","Data":"4ac4a65e17730cc898b7a1caa2cf6f59c848e5dbfb4544495d7d26d412afc515"} Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.417626 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.423451 4676 generic.go:334] "Generic (PLEG): container finished" podID="04db8ba1-c0de-4985-8a48-ead625786472" containerID="9f879edfb9e19e2ba5d1f7acc73cb74d671c4cd334c91721768ab3acc3bba0e9" exitCode=0 Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.423624 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-g5dnq" event={"ID":"04db8ba1-c0de-4985-8a48-ead625786472","Type":"ContainerDied","Data":"9f879edfb9e19e2ba5d1f7acc73cb74d671c4cd334c91721768ab3acc3bba0e9"} Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.435266 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-765f6cdf58-5q9v9" event={"ID":"21e63383-223a-4247-8589-03ab5a33f980","Type":"ContainerStarted","Data":"868be920f6388bdbc355ce503a74e95b8e95bf93569e1a5a274cfdb739ad2428"} Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.435495 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.435559 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-765f6cdf58-5q9v9" event={"ID":"21e63383-223a-4247-8589-03ab5a33f980","Type":"ContainerStarted","Data":"8c36a476f94bdc3802c17e235d4cf6c0cf6f6b6b85d7555584c90cce1257c52c"} Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.435627 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-765f6cdf58-5q9v9" event={"ID":"21e63383-223a-4247-8589-03ab5a33f980","Type":"ContainerStarted","Data":"76364b6f4d0e1a256cba5e253a99692a3f603a368a9595e44e963e8a0e1ff19c"} Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.435698 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.452994 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" podStartSLOduration=8.452979474 podStartE2EDuration="8.452979474s" podCreationTimestamp="2026-01-24 00:22:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:34.437919966 +0000 UTC m=+1138.467890967" watchObservedRunningTime="2026-01-24 00:22:34.452979474 +0000 UTC m=+1138.482950475" Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.504279 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-765f6cdf58-5q9v9" podStartSLOduration=4.504260964 podStartE2EDuration="4.504260964s" podCreationTimestamp="2026-01-24 00:22:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:34.498039414 +0000 UTC m=+1138.528010415" watchObservedRunningTime="2026-01-24 00:22:34.504260964 +0000 UTC m=+1138.534231965" Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.678540 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:34 crc kubenswrapper[4676]: I0124 00:22:34.922868 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.028440 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-scripts\") pod \"a7327fe7-179a-4492-8823-94b1067c17d4\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.028560 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-credential-keys\") pod \"a7327fe7-179a-4492-8823-94b1067c17d4\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.028620 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-fernet-keys\") pod \"a7327fe7-179a-4492-8823-94b1067c17d4\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.028672 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-config-data\") pod \"a7327fe7-179a-4492-8823-94b1067c17d4\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.028691 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjz2c\" (UniqueName: \"kubernetes.io/projected/a7327fe7-179a-4492-8823-94b1067c17d4-kube-api-access-wjz2c\") pod \"a7327fe7-179a-4492-8823-94b1067c17d4\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.028714 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-combined-ca-bundle\") pod \"a7327fe7-179a-4492-8823-94b1067c17d4\" (UID: \"a7327fe7-179a-4492-8823-94b1067c17d4\") " Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.033211 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-scripts" (OuterVolumeSpecName: "scripts") pod "a7327fe7-179a-4492-8823-94b1067c17d4" (UID: "a7327fe7-179a-4492-8823-94b1067c17d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.034599 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a7327fe7-179a-4492-8823-94b1067c17d4" (UID: "a7327fe7-179a-4492-8823-94b1067c17d4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.035295 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a7327fe7-179a-4492-8823-94b1067c17d4" (UID: "a7327fe7-179a-4492-8823-94b1067c17d4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.043497 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7327fe7-179a-4492-8823-94b1067c17d4-kube-api-access-wjz2c" (OuterVolumeSpecName: "kube-api-access-wjz2c") pod "a7327fe7-179a-4492-8823-94b1067c17d4" (UID: "a7327fe7-179a-4492-8823-94b1067c17d4"). InnerVolumeSpecName "kube-api-access-wjz2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.057541 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-config-data" (OuterVolumeSpecName: "config-data") pod "a7327fe7-179a-4492-8823-94b1067c17d4" (UID: "a7327fe7-179a-4492-8823-94b1067c17d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.057642 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7327fe7-179a-4492-8823-94b1067c17d4" (UID: "a7327fe7-179a-4492-8823-94b1067c17d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.131293 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjz2c\" (UniqueName: \"kubernetes.io/projected/a7327fe7-179a-4492-8823-94b1067c17d4-kube-api-access-wjz2c\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.131333 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.131342 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.131351 4676 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.131360 4676 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.131371 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7327fe7-179a-4492-8823-94b1067c17d4-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.457136 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-l49lm" event={"ID":"a7327fe7-179a-4492-8823-94b1067c17d4","Type":"ContainerDied","Data":"02e78984834a58ca6226d3263c50f11bea3fc65e0cf4f6d946dcccb6c1c331db"} Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.457411 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02e78984834a58ca6226d3263c50f11bea3fc65e0cf4f6d946dcccb6c1c331db" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.457847 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-l49lm" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.516183 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-849b597d57-kw79c"] Jan 24 00:22:35 crc kubenswrapper[4676]: E0124 00:22:35.516559 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7327fe7-179a-4492-8823-94b1067c17d4" containerName="keystone-bootstrap" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.516578 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7327fe7-179a-4492-8823-94b1067c17d4" containerName="keystone-bootstrap" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.516788 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7327fe7-179a-4492-8823-94b1067c17d4" containerName="keystone-bootstrap" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.517281 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.521546 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.521812 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.521980 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-fcgg9" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.522094 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.524452 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.524730 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.534006 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-849b597d57-kw79c"] Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.639708 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-public-tls-certs\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.639781 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-config-data\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.639801 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd9gh\" (UniqueName: \"kubernetes.io/projected/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-kube-api-access-rd9gh\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.639838 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-fernet-keys\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.639903 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-internal-tls-certs\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.639926 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-combined-ca-bundle\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.643779 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-credential-keys\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.643948 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-scripts\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.758862 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-scripts\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.758918 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-public-tls-certs\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.758949 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-config-data\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.758967 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd9gh\" (UniqueName: \"kubernetes.io/projected/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-kube-api-access-rd9gh\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.759000 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-fernet-keys\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.764919 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-internal-tls-certs\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.770111 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-combined-ca-bundle\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.770170 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-credential-keys\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.775773 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-scripts\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.776084 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-internal-tls-certs\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.776207 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-credential-keys\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.776861 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-fernet-keys\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.777985 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-public-tls-certs\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.779992 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-config-data\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.780948 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-combined-ca-bundle\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.792126 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd9gh\" (UniqueName: \"kubernetes.io/projected/29bf9e4b-4fb3-41f4-9280-f7ea2e61a844-kube-api-access-rd9gh\") pod \"keystone-849b597d57-kw79c\" (UID: \"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844\") " pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.834822 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:35 crc kubenswrapper[4676]: I0124 00:22:35.959162 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.004609 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qvjc\" (UniqueName: \"kubernetes.io/projected/04db8ba1-c0de-4985-8a48-ead625786472-kube-api-access-7qvjc\") pod \"04db8ba1-c0de-4985-8a48-ead625786472\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.004718 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-db-sync-config-data\") pod \"04db8ba1-c0de-4985-8a48-ead625786472\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.006477 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-combined-ca-bundle\") pod \"04db8ba1-c0de-4985-8a48-ead625786472\" (UID: \"04db8ba1-c0de-4985-8a48-ead625786472\") " Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.011802 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "04db8ba1-c0de-4985-8a48-ead625786472" (UID: "04db8ba1-c0de-4985-8a48-ead625786472"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.059183 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04db8ba1-c0de-4985-8a48-ead625786472-kube-api-access-7qvjc" (OuterVolumeSpecName: "kube-api-access-7qvjc") pod "04db8ba1-c0de-4985-8a48-ead625786472" (UID: "04db8ba1-c0de-4985-8a48-ead625786472"). InnerVolumeSpecName "kube-api-access-7qvjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.078163 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04db8ba1-c0de-4985-8a48-ead625786472" (UID: "04db8ba1-c0de-4985-8a48-ead625786472"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.110301 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qvjc\" (UniqueName: \"kubernetes.io/projected/04db8ba1-c0de-4985-8a48-ead625786472-kube-api-access-7qvjc\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.110336 4676 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.110347 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04db8ba1-c0de-4985-8a48-ead625786472-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.382715 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-849b597d57-kw79c"] Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.504763 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-g5dnq" event={"ID":"04db8ba1-c0de-4985-8a48-ead625786472","Type":"ContainerDied","Data":"534f710da5fb3c7907cdb2b5714b0d256a8306b5476b74f728dc51abaa10f9c0"} Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.504799 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="534f710da5fb3c7907cdb2b5714b0d256a8306b5476b74f728dc51abaa10f9c0" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.504858 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-g5dnq" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.522447 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-849b597d57-kw79c" event={"ID":"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844","Type":"ContainerStarted","Data":"9d7106c5a67270c418b542070c878ce61db9ee79ef8f26b080376fbfa65270da"} Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.737362 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-68bd7fb46c-sflbz"] Jan 24 00:22:36 crc kubenswrapper[4676]: E0124 00:22:36.737855 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04db8ba1-c0de-4985-8a48-ead625786472" containerName="barbican-db-sync" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.737870 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="04db8ba1-c0de-4985-8a48-ead625786472" containerName="barbican-db-sync" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.738048 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="04db8ba1-c0de-4985-8a48-ead625786472" containerName="barbican-db-sync" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.738980 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.742412 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-96w42" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.742616 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.742836 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.766594 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68bd7fb46c-sflbz"] Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.781764 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-79ccd69c74-nj8k6"] Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.783163 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.800904 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.813440 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79ccd69c74-nj8k6"] Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823227 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a05052ce-062e-423c-80cf-78349e42718f-config-data-custom\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823269 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/009f35e0-3a98-453b-b92e-8db9e5c92798-config-data-custom\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823305 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a05052ce-062e-423c-80cf-78349e42718f-logs\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823336 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxbwp\" (UniqueName: \"kubernetes.io/projected/a05052ce-062e-423c-80cf-78349e42718f-kube-api-access-mxbwp\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823354 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p982\" (UniqueName: \"kubernetes.io/projected/009f35e0-3a98-453b-b92e-8db9e5c92798-kube-api-access-2p982\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823395 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a05052ce-062e-423c-80cf-78349e42718f-combined-ca-bundle\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823415 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/009f35e0-3a98-453b-b92e-8db9e5c92798-combined-ca-bundle\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823444 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/009f35e0-3a98-453b-b92e-8db9e5c92798-config-data\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823461 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a05052ce-062e-423c-80cf-78349e42718f-config-data\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.823508 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/009f35e0-3a98-453b-b92e-8db9e5c92798-logs\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.899493 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-r22sn"] Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.900062 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" podUID="da4decd9-74a7-4128-a34b-e30de6318a09" containerName="dnsmasq-dns" containerID="cri-o://4ac4a65e17730cc898b7a1caa2cf6f59c848e5dbfb4544495d7d26d412afc515" gracePeriod=10 Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.927298 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/009f35e0-3a98-453b-b92e-8db9e5c92798-logs\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.927352 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a05052ce-062e-423c-80cf-78349e42718f-config-data-custom\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.927390 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/009f35e0-3a98-453b-b92e-8db9e5c92798-config-data-custom\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.927418 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a05052ce-062e-423c-80cf-78349e42718f-logs\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.927450 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxbwp\" (UniqueName: \"kubernetes.io/projected/a05052ce-062e-423c-80cf-78349e42718f-kube-api-access-mxbwp\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.931041 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p982\" (UniqueName: \"kubernetes.io/projected/009f35e0-3a98-453b-b92e-8db9e5c92798-kube-api-access-2p982\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.931099 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a05052ce-062e-423c-80cf-78349e42718f-combined-ca-bundle\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.931123 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/009f35e0-3a98-453b-b92e-8db9e5c92798-combined-ca-bundle\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.931172 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/009f35e0-3a98-453b-b92e-8db9e5c92798-config-data\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.931192 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a05052ce-062e-423c-80cf-78349e42718f-config-data\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.930751 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a05052ce-062e-423c-80cf-78349e42718f-logs\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.928520 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/009f35e0-3a98-453b-b92e-8db9e5c92798-logs\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.933256 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/009f35e0-3a98-453b-b92e-8db9e5c92798-config-data-custom\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.939097 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a05052ce-062e-423c-80cf-78349e42718f-combined-ca-bundle\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.943306 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a05052ce-062e-423c-80cf-78349e42718f-config-data\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.944870 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/009f35e0-3a98-453b-b92e-8db9e5c92798-combined-ca-bundle\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.950918 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xg9nt"] Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.952321 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a05052ce-062e-423c-80cf-78349e42718f-config-data-custom\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.952479 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.975425 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xg9nt"] Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.977288 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/009f35e0-3a98-453b-b92e-8db9e5c92798-config-data\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:36 crc kubenswrapper[4676]: I0124 00:22:36.981705 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p982\" (UniqueName: \"kubernetes.io/projected/009f35e0-3a98-453b-b92e-8db9e5c92798-kube-api-access-2p982\") pod \"barbican-worker-68bd7fb46c-sflbz\" (UID: \"009f35e0-3a98-453b-b92e-8db9e5c92798\") " pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.001414 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxbwp\" (UniqueName: \"kubernetes.io/projected/a05052ce-062e-423c-80cf-78349e42718f-kube-api-access-mxbwp\") pod \"barbican-keystone-listener-79ccd69c74-nj8k6\" (UID: \"a05052ce-062e-423c-80cf-78349e42718f\") " pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.033686 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.033804 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.033881 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-config\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.033927 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j5s9\" (UniqueName: \"kubernetes.io/projected/e1f791cc-7690-4f80-9007-a14fcca8a632-kube-api-access-9j5s9\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.036713 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.036771 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.085461 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.087529 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.087579 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.087589 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.158131 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-68bd7fb46c-sflbz" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.164951 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.165019 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.165061 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.165146 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.166470 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-config\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.166484 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.166561 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j5s9\" (UniqueName: \"kubernetes.io/projected/e1f791cc-7690-4f80-9007-a14fcca8a632-kube-api-access-9j5s9\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.183248 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.184610 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.185507 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-config\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.185671 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.200274 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.201213 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5d7c68bc86-pdds6"] Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.211631 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.214849 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.225949 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.227083 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j5s9\" (UniqueName: \"kubernetes.io/projected/e1f791cc-7690-4f80-9007-a14fcca8a632-kube-api-access-9j5s9\") pod \"dnsmasq-dns-848cf88cfc-xg9nt\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.249634 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.251704 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d7c68bc86-pdds6"] Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.395643 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data-custom\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.395741 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.395769 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-logs\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.395784 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfjfx\" (UniqueName: \"kubernetes.io/projected/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-kube-api-access-bfjfx\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.395798 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-combined-ca-bundle\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.487657 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.497878 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data-custom\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.497961 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.497988 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-logs\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.498003 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-combined-ca-bundle\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.498017 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfjfx\" (UniqueName: \"kubernetes.io/projected/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-kube-api-access-bfjfx\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.499024 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-logs\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.507539 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data-custom\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.507570 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.509276 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-combined-ca-bundle\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.521769 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfjfx\" (UniqueName: \"kubernetes.io/projected/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-kube-api-access-bfjfx\") pod \"barbican-api-5d7c68bc86-pdds6\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.542413 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.605901 4676 generic.go:334] "Generic (PLEG): container finished" podID="da4decd9-74a7-4128-a34b-e30de6318a09" containerID="4ac4a65e17730cc898b7a1caa2cf6f59c848e5dbfb4544495d7d26d412afc515" exitCode=0 Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.605963 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" event={"ID":"da4decd9-74a7-4128-a34b-e30de6318a09","Type":"ContainerDied","Data":"4ac4a65e17730cc898b7a1caa2cf6f59c848e5dbfb4544495d7d26d412afc515"} Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.649629 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-849b597d57-kw79c" event={"ID":"29bf9e4b-4fb3-41f4-9280-f7ea2e61a844","Type":"ContainerStarted","Data":"85f4314bb55ddae81a510b55326d40b4a0b3c015ae45e8ec0fc346374d088d81"} Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.649872 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.672166 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-849b597d57-kw79c" podStartSLOduration=2.672149998 podStartE2EDuration="2.672149998s" podCreationTimestamp="2026-01-24 00:22:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:37.67026375 +0000 UTC m=+1141.700234751" watchObservedRunningTime="2026-01-24 00:22:37.672149998 +0000 UTC m=+1141.702120999" Jan 24 00:22:37 crc kubenswrapper[4676]: I0124 00:22:37.914335 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.021620 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-swift-storage-0\") pod \"da4decd9-74a7-4128-a34b-e30de6318a09\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.024425 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-nb\") pod \"da4decd9-74a7-4128-a34b-e30de6318a09\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.024549 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-sb\") pod \"da4decd9-74a7-4128-a34b-e30de6318a09\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.024569 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-svc\") pod \"da4decd9-74a7-4128-a34b-e30de6318a09\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.024612 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq96k\" (UniqueName: \"kubernetes.io/projected/da4decd9-74a7-4128-a34b-e30de6318a09-kube-api-access-cq96k\") pod \"da4decd9-74a7-4128-a34b-e30de6318a09\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.024639 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-config\") pod \"da4decd9-74a7-4128-a34b-e30de6318a09\" (UID: \"da4decd9-74a7-4128-a34b-e30de6318a09\") " Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.079163 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da4decd9-74a7-4128-a34b-e30de6318a09-kube-api-access-cq96k" (OuterVolumeSpecName: "kube-api-access-cq96k") pod "da4decd9-74a7-4128-a34b-e30de6318a09" (UID: "da4decd9-74a7-4128-a34b-e30de6318a09"). InnerVolumeSpecName "kube-api-access-cq96k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.126402 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq96k\" (UniqueName: \"kubernetes.io/projected/da4decd9-74a7-4128-a34b-e30de6318a09-kube-api-access-cq96k\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.127432 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-68bd7fb46c-sflbz"] Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.276128 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "da4decd9-74a7-4128-a34b-e30de6318a09" (UID: "da4decd9-74a7-4128-a34b-e30de6318a09"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.289893 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "da4decd9-74a7-4128-a34b-e30de6318a09" (UID: "da4decd9-74a7-4128-a34b-e30de6318a09"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.308798 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.330247 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da4decd9-74a7-4128-a34b-e30de6318a09" (UID: "da4decd9-74a7-4128-a34b-e30de6318a09"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.339622 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.339661 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.339670 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.344883 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "da4decd9-74a7-4128-a34b-e30de6318a09" (UID: "da4decd9-74a7-4128-a34b-e30de6318a09"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.370167 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-config" (OuterVolumeSpecName: "config") pod "da4decd9-74a7-4128-a34b-e30de6318a09" (UID: "da4decd9-74a7-4128-a34b-e30de6318a09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.432413 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-79ccd69c74-nj8k6"] Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.440944 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.440969 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da4decd9-74a7-4128-a34b-e30de6318a09-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.614434 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xg9nt"] Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.623954 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d7c68bc86-pdds6"] Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.671398 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68bd7fb46c-sflbz" event={"ID":"009f35e0-3a98-453b-b92e-8db9e5c92798","Type":"ContainerStarted","Data":"68e1769fe551fc67b384bce7f563465fd1d599b8f8bb7ecda58660cc9b0ef80f"} Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.686718 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7c68bc86-pdds6" event={"ID":"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4","Type":"ContainerStarted","Data":"779fb9c258d57c3a4a4641f3663256b4984fdbbbea094e97da3fbf8189a0afcc"} Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.691909 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" event={"ID":"a05052ce-062e-423c-80cf-78349e42718f","Type":"ContainerStarted","Data":"09c7884475f21bcc53519656b4e2b5987f770f2355ebbed13f07205caaa7988b"} Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.702709 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.705503 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-r22sn" event={"ID":"da4decd9-74a7-4128-a34b-e30de6318a09","Type":"ContainerDied","Data":"a839f941eaa0eb3fef79a831fc52f7ec100168032ef563b034304acff5b0a94c"} Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.705565 4676 scope.go:117] "RemoveContainer" containerID="4ac4a65e17730cc898b7a1caa2cf6f59c848e5dbfb4544495d7d26d412afc515" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.742908 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" event={"ID":"e1f791cc-7690-4f80-9007-a14fcca8a632","Type":"ContainerStarted","Data":"2612852ab10c2bc7fb166122d80d7317364b784205eafc69c7981f567d09ca8e"} Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.853064 4676 scope.go:117] "RemoveContainer" containerID="8f47f3dbe46f5da8dd23954963ad8dc0cd2663f72dd53c26d9fe4db7d0341294" Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.898450 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-r22sn"] Jan 24 00:22:38 crc kubenswrapper[4676]: I0124 00:22:38.917429 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-r22sn"] Jan 24 00:22:39 crc kubenswrapper[4676]: I0124 00:22:39.380119 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:22:39 crc kubenswrapper[4676]: I0124 00:22:39.380343 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:22:39 crc kubenswrapper[4676]: I0124 00:22:39.819754 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7c68bc86-pdds6" event={"ID":"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4","Type":"ContainerStarted","Data":"dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92"} Jan 24 00:22:39 crc kubenswrapper[4676]: I0124 00:22:39.820026 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7c68bc86-pdds6" event={"ID":"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4","Type":"ContainerStarted","Data":"f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f"} Jan 24 00:22:39 crc kubenswrapper[4676]: I0124 00:22:39.821076 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:39 crc kubenswrapper[4676]: I0124 00:22:39.821184 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:39 crc kubenswrapper[4676]: I0124 00:22:39.826194 4676 generic.go:334] "Generic (PLEG): container finished" podID="e1f791cc-7690-4f80-9007-a14fcca8a632" containerID="cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a" exitCode=0 Jan 24 00:22:39 crc kubenswrapper[4676]: I0124 00:22:39.826265 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" event={"ID":"e1f791cc-7690-4f80-9007-a14fcca8a632","Type":"ContainerDied","Data":"cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a"} Jan 24 00:22:39 crc kubenswrapper[4676]: I0124 00:22:39.911343 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5d7c68bc86-pdds6" podStartSLOduration=2.911324877 podStartE2EDuration="2.911324877s" podCreationTimestamp="2026-01-24 00:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:39.883756469 +0000 UTC m=+1143.913727470" watchObservedRunningTime="2026-01-24 00:22:39.911324877 +0000 UTC m=+1143.941295878" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.170804 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-54b7b855f4-s49zw"] Jan 24 00:22:40 crc kubenswrapper[4676]: E0124 00:22:40.171245 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da4decd9-74a7-4128-a34b-e30de6318a09" containerName="init" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.171266 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="da4decd9-74a7-4128-a34b-e30de6318a09" containerName="init" Jan 24 00:22:40 crc kubenswrapper[4676]: E0124 00:22:40.171286 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da4decd9-74a7-4128-a34b-e30de6318a09" containerName="dnsmasq-dns" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.171294 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="da4decd9-74a7-4128-a34b-e30de6318a09" containerName="dnsmasq-dns" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.171539 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="da4decd9-74a7-4128-a34b-e30de6318a09" containerName="dnsmasq-dns" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.172446 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.173340 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.176074 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54b7b855f4-s49zw"] Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.181856 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.182090 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.281333 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da4decd9-74a7-4128-a34b-e30de6318a09" path="/var/lib/kubelet/pods/da4decd9-74a7-4128-a34b-e30de6318a09/volumes" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.310451 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-combined-ca-bundle\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.310510 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-public-tls-certs\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.310543 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5vkx\" (UniqueName: \"kubernetes.io/projected/c015646b-9361-4b5c-b465-a66d7fc5cc53-kube-api-access-n5vkx\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.310571 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-config-data-custom\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.310595 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c015646b-9361-4b5c-b465-a66d7fc5cc53-logs\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.310622 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-config-data\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.310654 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-internal-tls-certs\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.324428 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f876ddf46-fs7qv" podUID="ac7dce6b-3bd9-4ad9-9485-83d9384b8bad" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.412110 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c015646b-9361-4b5c-b465-a66d7fc5cc53-logs\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.412176 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-config-data\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.412216 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-internal-tls-certs\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.412275 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-combined-ca-bundle\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.412592 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c015646b-9361-4b5c-b465-a66d7fc5cc53-logs\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.413136 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-public-tls-certs\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.413174 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5vkx\" (UniqueName: \"kubernetes.io/projected/c015646b-9361-4b5c-b465-a66d7fc5cc53-kube-api-access-n5vkx\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.413201 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-config-data-custom\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.418193 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-combined-ca-bundle\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.418487 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-config-data\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.421910 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-internal-tls-certs\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.422337 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-config-data-custom\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.425494 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c015646b-9361-4b5c-b465-a66d7fc5cc53-public-tls-certs\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.436758 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5vkx\" (UniqueName: \"kubernetes.io/projected/c015646b-9361-4b5c-b465-a66d7fc5cc53-kube-api-access-n5vkx\") pod \"barbican-api-54b7b855f4-s49zw\" (UID: \"c015646b-9361-4b5c-b465-a66d7fc5cc53\") " pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.494114 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.896187 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" event={"ID":"e1f791cc-7690-4f80-9007-a14fcca8a632","Type":"ContainerStarted","Data":"e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b"} Jan 24 00:22:40 crc kubenswrapper[4676]: I0124 00:22:40.925042 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" podStartSLOduration=4.925024486 podStartE2EDuration="4.925024486s" podCreationTimestamp="2026-01-24 00:22:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:40.919341363 +0000 UTC m=+1144.949312364" watchObservedRunningTime="2026-01-24 00:22:40.925024486 +0000 UTC m=+1144.954995477" Jan 24 00:22:41 crc kubenswrapper[4676]: I0124 00:22:41.078247 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-54b7b855f4-s49zw"] Jan 24 00:22:41 crc kubenswrapper[4676]: I0124 00:22:41.907585 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7klnv" event={"ID":"d1163cc6-7ce1-4f2b-9d4a-f3e215177842","Type":"ContainerStarted","Data":"e4299d6b00616548489e178e72141c6cc9961df799e192961e6eb82793cfee65"} Jan 24 00:22:41 crc kubenswrapper[4676]: I0124 00:22:41.908895 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:41 crc kubenswrapper[4676]: I0124 00:22:41.927056 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7klnv" podStartSLOduration=5.797360287 podStartE2EDuration="51.92703827s" podCreationTimestamp="2026-01-24 00:21:50 +0000 UTC" firstStartedPulling="2026-01-24 00:21:53.684897811 +0000 UTC m=+1097.714868812" lastFinishedPulling="2026-01-24 00:22:39.814575804 +0000 UTC m=+1143.844546795" observedRunningTime="2026-01-24 00:22:41.921494611 +0000 UTC m=+1145.951465612" watchObservedRunningTime="2026-01-24 00:22:41.92703827 +0000 UTC m=+1145.957009271" Jan 24 00:22:41 crc kubenswrapper[4676]: I0124 00:22:41.998851 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 24 00:22:41 crc kubenswrapper[4676]: I0124 00:22:41.998940 4676 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 24 00:22:42 crc kubenswrapper[4676]: I0124 00:22:42.659673 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 24 00:22:42 crc kubenswrapper[4676]: I0124 00:22:42.926801 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54b7b855f4-s49zw" event={"ID":"c015646b-9361-4b5c-b465-a66d7fc5cc53","Type":"ContainerStarted","Data":"e496978521b70573fd52b3add57183ce8b6b6e0787cf2107446d0ef525aff3d2"} Jan 24 00:22:42 crc kubenswrapper[4676]: I0124 00:22:42.927871 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54b7b855f4-s49zw" event={"ID":"c015646b-9361-4b5c-b465-a66d7fc5cc53","Type":"ContainerStarted","Data":"d0be057cf18c520ed2842a7fe9c36dc7e6de7a1d0254d9e200d5899fbc1c27d5"} Jan 24 00:22:42 crc kubenswrapper[4676]: I0124 00:22:42.929738 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68bd7fb46c-sflbz" event={"ID":"009f35e0-3a98-453b-b92e-8db9e5c92798","Type":"ContainerStarted","Data":"269c96192fce502f2b814bdc98b7995edf5b48d9aa750af264d7848e970f68c6"} Jan 24 00:22:42 crc kubenswrapper[4676]: I0124 00:22:42.956279 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" event={"ID":"a05052ce-062e-423c-80cf-78349e42718f","Type":"ContainerStarted","Data":"f06df1db0559b1557e08daa5944cfb962f47144c86003fc9692dd66eaa9ee71e"} Jan 24 00:22:43 crc kubenswrapper[4676]: I0124 00:22:43.966844 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-54b7b855f4-s49zw" event={"ID":"c015646b-9361-4b5c-b465-a66d7fc5cc53","Type":"ContainerStarted","Data":"f0bb652a2091409660eb1ee979e75bc34ed86c2ef99c95a31fdf1e84f770ce3a"} Jan 24 00:22:43 crc kubenswrapper[4676]: I0124 00:22:43.967139 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:43 crc kubenswrapper[4676]: I0124 00:22:43.967151 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:43 crc kubenswrapper[4676]: I0124 00:22:43.974429 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-68bd7fb46c-sflbz" event={"ID":"009f35e0-3a98-453b-b92e-8db9e5c92798","Type":"ContainerStarted","Data":"b7aa88ef8c3f568bb1563d37814bc2675d26dd45c4f8e4f2822b1fc05113cf2a"} Jan 24 00:22:43 crc kubenswrapper[4676]: I0124 00:22:43.990329 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" event={"ID":"a05052ce-062e-423c-80cf-78349e42718f","Type":"ContainerStarted","Data":"a0cda41f1acb5a3f79a1ec315685e0142319a74b0456f90b43600bdc9997e6f3"} Jan 24 00:22:43 crc kubenswrapper[4676]: I0124 00:22:43.995921 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-54b7b855f4-s49zw" podStartSLOduration=3.995902549 podStartE2EDuration="3.995902549s" podCreationTimestamp="2026-01-24 00:22:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:43.988815203 +0000 UTC m=+1148.018786204" watchObservedRunningTime="2026-01-24 00:22:43.995902549 +0000 UTC m=+1148.025873550" Jan 24 00:22:44 crc kubenswrapper[4676]: I0124 00:22:44.022721 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-68bd7fb46c-sflbz" podStartSLOduration=3.956530082 podStartE2EDuration="8.022702284s" podCreationTimestamp="2026-01-24 00:22:36 +0000 UTC" firstStartedPulling="2026-01-24 00:22:38.206891001 +0000 UTC m=+1142.236862002" lastFinishedPulling="2026-01-24 00:22:42.273063203 +0000 UTC m=+1146.303034204" observedRunningTime="2026-01-24 00:22:44.012730001 +0000 UTC m=+1148.042701012" watchObservedRunningTime="2026-01-24 00:22:44.022702284 +0000 UTC m=+1148.052673285" Jan 24 00:22:47 crc kubenswrapper[4676]: I0124 00:22:47.491585 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:22:47 crc kubenswrapper[4676]: I0124 00:22:47.515055 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-79ccd69c74-nj8k6" podStartSLOduration=7.682827598 podStartE2EDuration="11.515039755s" podCreationTimestamp="2026-01-24 00:22:36 +0000 UTC" firstStartedPulling="2026-01-24 00:22:38.443853327 +0000 UTC m=+1142.473824328" lastFinishedPulling="2026-01-24 00:22:42.276065494 +0000 UTC m=+1146.306036485" observedRunningTime="2026-01-24 00:22:44.042493036 +0000 UTC m=+1148.072464037" watchObservedRunningTime="2026-01-24 00:22:47.515039755 +0000 UTC m=+1151.545010756" Jan 24 00:22:47 crc kubenswrapper[4676]: I0124 00:22:47.549687 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-hpwcp"] Jan 24 00:22:47 crc kubenswrapper[4676]: I0124 00:22:47.549924 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" podUID="2417692e-7845-446e-a60a-bb3842851912" containerName="dnsmasq-dns" containerID="cri-o://d6cc348312351f0aa6a3a89dece34eea57cb75b4f69f18c04d668c9984cdefc4" gracePeriod=10 Jan 24 00:22:48 crc kubenswrapper[4676]: I0124 00:22:48.025493 4676 generic.go:334] "Generic (PLEG): container finished" podID="2417692e-7845-446e-a60a-bb3842851912" containerID="d6cc348312351f0aa6a3a89dece34eea57cb75b4f69f18c04d668c9984cdefc4" exitCode=0 Jan 24 00:22:48 crc kubenswrapper[4676]: I0124 00:22:48.025570 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" event={"ID":"2417692e-7845-446e-a60a-bb3842851912","Type":"ContainerDied","Data":"d6cc348312351f0aa6a3a89dece34eea57cb75b4f69f18c04d668c9984cdefc4"} Jan 24 00:22:49 crc kubenswrapper[4676]: I0124 00:22:49.034011 4676 generic.go:334] "Generic (PLEG): container finished" podID="d1163cc6-7ce1-4f2b-9d4a-f3e215177842" containerID="e4299d6b00616548489e178e72141c6cc9961df799e192961e6eb82793cfee65" exitCode=0 Jan 24 00:22:49 crc kubenswrapper[4676]: I0124 00:22:49.034217 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7klnv" event={"ID":"d1163cc6-7ce1-4f2b-9d4a-f3e215177842","Type":"ContainerDied","Data":"e4299d6b00616548489e178e72141c6cc9961df799e192961e6eb82793cfee65"} Jan 24 00:22:49 crc kubenswrapper[4676]: I0124 00:22:49.734692 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.178673 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.322002 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f876ddf46-fs7qv" podUID="ac7dce6b-3bd9-4ad9-9485-83d9384b8bad" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.864544 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.910220 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-sb\") pod \"2417692e-7845-446e-a60a-bb3842851912\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.910257 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rjk5\" (UniqueName: \"kubernetes.io/projected/2417692e-7845-446e-a60a-bb3842851912-kube-api-access-5rjk5\") pod \"2417692e-7845-446e-a60a-bb3842851912\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.910290 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-swift-storage-0\") pod \"2417692e-7845-446e-a60a-bb3842851912\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.910329 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-config\") pod \"2417692e-7845-446e-a60a-bb3842851912\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.910370 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-nb\") pod \"2417692e-7845-446e-a60a-bb3842851912\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.910527 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-svc\") pod \"2417692e-7845-446e-a60a-bb3842851912\" (UID: \"2417692e-7845-446e-a60a-bb3842851912\") " Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.922104 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7klnv" Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.939480 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2417692e-7845-446e-a60a-bb3842851912-kube-api-access-5rjk5" (OuterVolumeSpecName: "kube-api-access-5rjk5") pod "2417692e-7845-446e-a60a-bb3842851912" (UID: "2417692e-7845-446e-a60a-bb3842851912"). InnerVolumeSpecName "kube-api-access-5rjk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.965068 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2417692e-7845-446e-a60a-bb3842851912" (UID: "2417692e-7845-446e-a60a-bb3842851912"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:50 crc kubenswrapper[4676]: I0124 00:22:50.992821 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2417692e-7845-446e-a60a-bb3842851912" (UID: "2417692e-7845-446e-a60a-bb3842851912"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.012947 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-scripts\") pod \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.013005 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsgpg\" (UniqueName: \"kubernetes.io/projected/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-kube-api-access-rsgpg\") pod \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.013099 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-db-sync-config-data\") pod \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.013134 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-config-data\") pod \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.013178 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-combined-ca-bundle\") pod \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.013227 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-etc-machine-id\") pod \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\" (UID: \"d1163cc6-7ce1-4f2b-9d4a-f3e215177842\") " Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.013618 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.013634 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.013644 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rjk5\" (UniqueName: \"kubernetes.io/projected/2417692e-7845-446e-a60a-bb3842851912-kube-api-access-5rjk5\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.013676 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d1163cc6-7ce1-4f2b-9d4a-f3e215177842" (UID: "d1163cc6-7ce1-4f2b-9d4a-f3e215177842"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.042802 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d1163cc6-7ce1-4f2b-9d4a-f3e215177842" (UID: "d1163cc6-7ce1-4f2b-9d4a-f3e215177842"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.092004 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7klnv" event={"ID":"d1163cc6-7ce1-4f2b-9d4a-f3e215177842","Type":"ContainerDied","Data":"70e37739b2a173b291b1d289a2d2839fa4f44eb620a1fa26acee32cd0b9f1716"} Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.092055 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70e37739b2a173b291b1d289a2d2839fa4f44eb620a1fa26acee32cd0b9f1716" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.092154 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7klnv" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.105784 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-scripts" (OuterVolumeSpecName: "scripts") pod "d1163cc6-7ce1-4f2b-9d4a-f3e215177842" (UID: "d1163cc6-7ce1-4f2b-9d4a-f3e215177842"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.121656 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" event={"ID":"2417692e-7845-446e-a60a-bb3842851912","Type":"ContainerDied","Data":"a17194b7f35eb9f8070fa9acebc74f7ada011686f763d785550450b94c66597c"} Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.121714 4676 scope.go:117] "RemoveContainer" containerID="d6cc348312351f0aa6a3a89dece34eea57cb75b4f69f18c04d668c9984cdefc4" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.121860 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-hpwcp" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.128046 4676 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.128075 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.128084 4676 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.150652 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-kube-api-access-rsgpg" (OuterVolumeSpecName: "kube-api-access-rsgpg") pod "d1163cc6-7ce1-4f2b-9d4a-f3e215177842" (UID: "d1163cc6-7ce1-4f2b-9d4a-f3e215177842"). InnerVolumeSpecName "kube-api-access-rsgpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.161509 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2417692e-7845-446e-a60a-bb3842851912" (UID: "2417692e-7845-446e-a60a-bb3842851912"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.162665 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-config" (OuterVolumeSpecName: "config") pod "2417692e-7845-446e-a60a-bb3842851912" (UID: "2417692e-7845-446e-a60a-bb3842851912"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.170920 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2417692e-7845-446e-a60a-bb3842851912" (UID: "2417692e-7845-446e-a60a-bb3842851912"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.172661 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1163cc6-7ce1-4f2b-9d4a-f3e215177842" (UID: "d1163cc6-7ce1-4f2b-9d4a-f3e215177842"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.242329 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-config-data" (OuterVolumeSpecName: "config-data") pod "d1163cc6-7ce1-4f2b-9d4a-f3e215177842" (UID: "d1163cc6-7ce1-4f2b-9d4a-f3e215177842"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.243760 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.243779 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.243793 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsgpg\" (UniqueName: \"kubernetes.io/projected/d1163cc6-7ce1-4f2b-9d4a-f3e215177842-kube-api-access-rsgpg\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.243802 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.243811 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.243820 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2417692e-7845-446e-a60a-bb3842851912-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.268467 4676 scope.go:117] "RemoveContainer" containerID="4c0f072637105b041cb28fbed44761b77e903f1e8315f62df0bde2e50a63a7fa" Jan 24 00:22:51 crc kubenswrapper[4676]: E0124 00:22:51.354832 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.469286 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 24 00:22:51 crc kubenswrapper[4676]: E0124 00:22:51.469633 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2417692e-7845-446e-a60a-bb3842851912" containerName="dnsmasq-dns" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.469648 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2417692e-7845-446e-a60a-bb3842851912" containerName="dnsmasq-dns" Jan 24 00:22:51 crc kubenswrapper[4676]: E0124 00:22:51.469662 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2417692e-7845-446e-a60a-bb3842851912" containerName="init" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.469669 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2417692e-7845-446e-a60a-bb3842851912" containerName="init" Jan 24 00:22:51 crc kubenswrapper[4676]: E0124 00:22:51.469708 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1163cc6-7ce1-4f2b-9d4a-f3e215177842" containerName="cinder-db-sync" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.469714 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1163cc6-7ce1-4f2b-9d4a-f3e215177842" containerName="cinder-db-sync" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.469895 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2417692e-7845-446e-a60a-bb3842851912" containerName="dnsmasq-dns" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.469917 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1163cc6-7ce1-4f2b-9d4a-f3e215177842" containerName="cinder-db-sync" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.472883 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.486962 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-vxll9" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.487048 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.487172 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.487299 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.487845 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-hpwcp"] Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.502443 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-hpwcp"] Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.524497 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.549327 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.549389 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-scripts\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.549430 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vknvh\" (UniqueName: \"kubernetes.io/projected/fb66b31d-c165-4930-b633-79707cbf22bb-kube-api-access-vknvh\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.549497 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.549519 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.549542 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb66b31d-c165-4930-b633-79707cbf22bb-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.584266 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-jbptt"] Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.600926 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.646028 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-jbptt"] Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652422 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-svc\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652479 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652523 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-scripts\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652546 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-config\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652563 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652604 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vknvh\" (UniqueName: \"kubernetes.io/projected/fb66b31d-c165-4930-b633-79707cbf22bb-kube-api-access-vknvh\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652650 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxx64\" (UniqueName: \"kubernetes.io/projected/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-kube-api-access-xxx64\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652676 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652692 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652734 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652755 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652778 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb66b31d-c165-4930-b633-79707cbf22bb-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.652860 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb66b31d-c165-4930-b633-79707cbf22bb-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.674112 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.676004 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.684875 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vknvh\" (UniqueName: \"kubernetes.io/projected/fb66b31d-c165-4930-b633-79707cbf22bb-kube-api-access-vknvh\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.693594 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.702667 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-scripts\") pod \"cinder-scheduler-0\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.734420 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.736014 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.747705 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.754143 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxx64\" (UniqueName: \"kubernetes.io/projected/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-kube-api-access-xxx64\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.754187 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.754213 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.754315 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-svc\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.754359 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-config\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.754415 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.755315 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.755425 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-svc\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.757420 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-config\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.758278 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.758467 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.771202 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.798663 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.817548 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxx64\" (UniqueName: \"kubernetes.io/projected/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-kube-api-access-xxx64\") pod \"dnsmasq-dns-6578955fd5-jbptt\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.857431 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-scripts\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.857668 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data-custom\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.857748 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbnnh\" (UniqueName: \"kubernetes.io/projected/a658cfcc-7de1-467d-8195-21f54cfaaf18-kube-api-access-hbnnh\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.857858 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.857979 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a658cfcc-7de1-467d-8195-21f54cfaaf18-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.858054 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a658cfcc-7de1-467d-8195-21f54cfaaf18-logs\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.858134 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.921509 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.963252 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a658cfcc-7de1-467d-8195-21f54cfaaf18-logs\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.963533 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.963575 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-scripts\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.963591 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data-custom\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.963606 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbnnh\" (UniqueName: \"kubernetes.io/projected/a658cfcc-7de1-467d-8195-21f54cfaaf18-kube-api-access-hbnnh\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.963657 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.963722 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a658cfcc-7de1-467d-8195-21f54cfaaf18-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.963793 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a658cfcc-7de1-467d-8195-21f54cfaaf18-etc-machine-id\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.963939 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a658cfcc-7de1-467d-8195-21f54cfaaf18-logs\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.982580 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-scripts\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.988224 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data-custom\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:51 crc kubenswrapper[4676]: I0124 00:22:51.995246 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.006890 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.026924 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbnnh\" (UniqueName: \"kubernetes.io/projected/a658cfcc-7de1-467d-8195-21f54cfaaf18-kube-api-access-hbnnh\") pod \"cinder-api-0\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " pod="openstack/cinder-api-0" Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.138626 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d56910-d4b7-45b1-8699-5af7eaa29b96","Type":"ContainerStarted","Data":"1b26cefde45473bb50d47ba50a2ec78ece32f247247ca63eb469045a70b4a673"} Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.138775 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="ceilometer-notification-agent" containerID="cri-o://f53a3c895796c44b94d61775eb726821f040873261a1a8fbdb79f578ca085ec5" gracePeriod=30 Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.139607 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.139891 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="proxy-httpd" containerID="cri-o://1b26cefde45473bb50d47ba50a2ec78ece32f247247ca63eb469045a70b4a673" gracePeriod=30 Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.139937 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="sg-core" containerID="cri-o://dae61f7a8060ccdb486fd53e86c1fca33bbb71e0fb9f0d456ad2e0a5a8007cbd" gracePeriod=30 Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.227922 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.275321 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2417692e-7845-446e-a60a-bb3842851912" path="/var/lib/kubelet/pods/2417692e-7845-446e-a60a-bb3842851912/volumes" Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.588597 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5d7c68bc86-pdds6" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.615984 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 24 00:22:52 crc kubenswrapper[4676]: I0124 00:22:52.704877 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-jbptt"] Jan 24 00:22:53 crc kubenswrapper[4676]: I0124 00:22:53.145899 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 24 00:22:53 crc kubenswrapper[4676]: I0124 00:22:53.151249 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" event={"ID":"9b4acc7d-6a5f-4544-b73e-61b9a620c17d","Type":"ContainerStarted","Data":"eec257beeff7ba3f585f1d2f0b787abd4407c78ad654a5cbdd01eb786faa02e2"} Jan 24 00:22:53 crc kubenswrapper[4676]: I0124 00:22:53.172496 4676 generic.go:334] "Generic (PLEG): container finished" podID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerID="dae61f7a8060ccdb486fd53e86c1fca33bbb71e0fb9f0d456ad2e0a5a8007cbd" exitCode=2 Jan 24 00:22:53 crc kubenswrapper[4676]: I0124 00:22:53.172536 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d56910-d4b7-45b1-8699-5af7eaa29b96","Type":"ContainerDied","Data":"dae61f7a8060ccdb486fd53e86c1fca33bbb71e0fb9f0d456ad2e0a5a8007cbd"} Jan 24 00:22:53 crc kubenswrapper[4676]: I0124 00:22:53.174262 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fb66b31d-c165-4930-b633-79707cbf22bb","Type":"ContainerStarted","Data":"f56885c2863c57838bc58c1bd3891e3e64ca0124e390aeca0b9564b1dc1517dd"} Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.279954 4676 generic.go:334] "Generic (PLEG): container finished" podID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerID="f53a3c895796c44b94d61775eb726821f040873261a1a8fbdb79f578ca085ec5" exitCode=0 Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.280570 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d56910-d4b7-45b1-8699-5af7eaa29b96","Type":"ContainerDied","Data":"f53a3c895796c44b94d61775eb726821f040873261a1a8fbdb79f578ca085ec5"} Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.298474 4676 generic.go:334] "Generic (PLEG): container finished" podID="9b4acc7d-6a5f-4544-b73e-61b9a620c17d" containerID="1da12014ab624fa95aa24ddc28c008acbe5db6c02d4d9e96a348452a0b9b9e38" exitCode=0 Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.298536 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" event={"ID":"9b4acc7d-6a5f-4544-b73e-61b9a620c17d","Type":"ContainerDied","Data":"1da12014ab624fa95aa24ddc28c008acbe5db6c02d4d9e96a348452a0b9b9e38"} Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.310568 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a658cfcc-7de1-467d-8195-21f54cfaaf18","Type":"ContainerStarted","Data":"102eb5e37f96d7d93fb2c20897bc25ad810e3aed4efdc7876fc59651cbc8312e"} Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.310633 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a658cfcc-7de1-467d-8195-21f54cfaaf18","Type":"ContainerStarted","Data":"5f5f73538e6873cbcb22a131253b622d3c4660ae4fc5a284910827195368977f"} Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.502971 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-54b7b855f4-s49zw" podUID="c015646b-9361-4b5c-b465-a66d7fc5cc53" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.550582 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.586612 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-5d7c68bc86-pdds6" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 24 00:22:54 crc kubenswrapper[4676]: I0124 00:22:54.871266 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:55 crc kubenswrapper[4676]: I0124 00:22:55.325603 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fb66b31d-c165-4930-b633-79707cbf22bb","Type":"ContainerStarted","Data":"617a7bf60c86bd8fb95135f8ae3c47f3775cca4441a612fa0eecdd996ac43e81"} Jan 24 00:22:55 crc kubenswrapper[4676]: I0124 00:22:55.354473 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" event={"ID":"9b4acc7d-6a5f-4544-b73e-61b9a620c17d","Type":"ContainerStarted","Data":"3c1cdcc2ffd783555c7e0583ecb6ac91d2d0c15034969f7f5f6f8dba79f95867"} Jan 24 00:22:55 crc kubenswrapper[4676]: I0124 00:22:55.355531 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:22:55 crc kubenswrapper[4676]: I0124 00:22:55.389304 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" podStartSLOduration=4.389283621 podStartE2EDuration="4.389283621s" podCreationTimestamp="2026-01-24 00:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:55.381286978 +0000 UTC m=+1159.411257979" watchObservedRunningTime="2026-01-24 00:22:55.389283621 +0000 UTC m=+1159.419254622" Jan 24 00:22:55 crc kubenswrapper[4676]: I0124 00:22:55.527626 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-54b7b855f4-s49zw" podUID="c015646b-9361-4b5c-b465-a66d7fc5cc53" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.353542 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.376496 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fb66b31d-c165-4930-b633-79707cbf22bb","Type":"ContainerStarted","Data":"2401fa1f3b792936f62c0e8fef2c85d82ca1c01d1b0b3520974286bb7dbc9b75"} Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.379558 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerName="cinder-api-log" containerID="cri-o://102eb5e37f96d7d93fb2c20897bc25ad810e3aed4efdc7876fc59651cbc8312e" gracePeriod=30 Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.379571 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a658cfcc-7de1-467d-8195-21f54cfaaf18","Type":"ContainerStarted","Data":"58af4da8742433d2dc7484bcefd21745ff4c7fa2b61dd06f0a0fc41c705c834a"} Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.379633 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerName="cinder-api" containerID="cri-o://58af4da8742433d2dc7484bcefd21745ff4c7fa2b61dd06f0a0fc41c705c834a" gracePeriod=30 Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.379682 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.448279 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.448261608 podStartE2EDuration="5.448261608s" podCreationTimestamp="2026-01-24 00:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:56.430970242 +0000 UTC m=+1160.460941243" watchObservedRunningTime="2026-01-24 00:22:56.448261608 +0000 UTC m=+1160.478232609" Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.483741 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.595564254 podStartE2EDuration="5.483724106s" podCreationTimestamp="2026-01-24 00:22:51 +0000 UTC" firstStartedPulling="2026-01-24 00:22:52.662616006 +0000 UTC m=+1156.692586997" lastFinishedPulling="2026-01-24 00:22:53.550775848 +0000 UTC m=+1157.580746849" observedRunningTime="2026-01-24 00:22:56.478823387 +0000 UTC m=+1160.508794388" watchObservedRunningTime="2026-01-24 00:22:56.483724106 +0000 UTC m=+1160.513695107" Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.680927 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.800391 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.926320 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b7b989687-7nkv4"] Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.926549 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b7b989687-7nkv4" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-api" containerID="cri-o://bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724" gracePeriod=30 Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.926769 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b7b989687-7nkv4" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-httpd" containerID="cri-o://b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae" gracePeriod=30 Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.944404 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-b7b989687-7nkv4" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.155:9696/\": read tcp 10.217.0.2:52290->10.217.0.155:9696: read: connection reset by peer" Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.969094 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bd579cfd9-q4npp"] Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.970912 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:56 crc kubenswrapper[4676]: I0124 00:22:56.984104 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bd579cfd9-q4npp"] Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.128834 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-internal-tls-certs\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.128898 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-public-tls-certs\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.128921 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-674zd\" (UniqueName: \"kubernetes.io/projected/6dda8455-6777-4efc-abf3-df547cf58339-kube-api-access-674zd\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.129138 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-config\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.129274 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-ovndb-tls-certs\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.129536 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-httpd-config\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.129585 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-combined-ca-bundle\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.231144 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-config\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.231202 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-ovndb-tls-certs\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.231268 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-httpd-config\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.231285 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-combined-ca-bundle\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.231331 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-internal-tls-certs\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.231361 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-public-tls-certs\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.231390 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-674zd\" (UniqueName: \"kubernetes.io/projected/6dda8455-6777-4efc-abf3-df547cf58339-kube-api-access-674zd\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.238219 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-public-tls-certs\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.238834 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-httpd-config\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.240708 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-combined-ca-bundle\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.241726 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-config\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.242765 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-internal-tls-certs\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.261202 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dda8455-6777-4efc-abf3-df547cf58339-ovndb-tls-certs\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.290610 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-674zd\" (UniqueName: \"kubernetes.io/projected/6dda8455-6777-4efc-abf3-df547cf58339-kube-api-access-674zd\") pod \"neutron-bd579cfd9-q4npp\" (UID: \"6dda8455-6777-4efc-abf3-df547cf58339\") " pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.298866 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.436308 4676 generic.go:334] "Generic (PLEG): container finished" podID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerID="58af4da8742433d2dc7484bcefd21745ff4c7fa2b61dd06f0a0fc41c705c834a" exitCode=0 Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.436571 4676 generic.go:334] "Generic (PLEG): container finished" podID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerID="102eb5e37f96d7d93fb2c20897bc25ad810e3aed4efdc7876fc59651cbc8312e" exitCode=143 Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.436610 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a658cfcc-7de1-467d-8195-21f54cfaaf18","Type":"ContainerDied","Data":"58af4da8742433d2dc7484bcefd21745ff4c7fa2b61dd06f0a0fc41c705c834a"} Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.436636 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a658cfcc-7de1-467d-8195-21f54cfaaf18","Type":"ContainerDied","Data":"102eb5e37f96d7d93fb2c20897bc25ad810e3aed4efdc7876fc59651cbc8312e"} Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.449838 4676 generic.go:334] "Generic (PLEG): container finished" podID="2e3ec968-d892-487a-930a-44c79123d54b" containerID="b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae" exitCode=0 Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.449906 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b7b989687-7nkv4" event={"ID":"2e3ec968-d892-487a-930a-44c79123d54b","Type":"ContainerDied","Data":"b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae"} Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.856425 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.954983 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a658cfcc-7de1-467d-8195-21f54cfaaf18-logs\") pod \"a658cfcc-7de1-467d-8195-21f54cfaaf18\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.955030 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-scripts\") pod \"a658cfcc-7de1-467d-8195-21f54cfaaf18\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.955066 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbnnh\" (UniqueName: \"kubernetes.io/projected/a658cfcc-7de1-467d-8195-21f54cfaaf18-kube-api-access-hbnnh\") pod \"a658cfcc-7de1-467d-8195-21f54cfaaf18\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.955100 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data-custom\") pod \"a658cfcc-7de1-467d-8195-21f54cfaaf18\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.955150 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data\") pod \"a658cfcc-7de1-467d-8195-21f54cfaaf18\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.955169 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a658cfcc-7de1-467d-8195-21f54cfaaf18-etc-machine-id\") pod \"a658cfcc-7de1-467d-8195-21f54cfaaf18\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.955276 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-combined-ca-bundle\") pod \"a658cfcc-7de1-467d-8195-21f54cfaaf18\" (UID: \"a658cfcc-7de1-467d-8195-21f54cfaaf18\") " Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.955365 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a658cfcc-7de1-467d-8195-21f54cfaaf18-logs" (OuterVolumeSpecName: "logs") pod "a658cfcc-7de1-467d-8195-21f54cfaaf18" (UID: "a658cfcc-7de1-467d-8195-21f54cfaaf18"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.955619 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a658cfcc-7de1-467d-8195-21f54cfaaf18-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.956847 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a658cfcc-7de1-467d-8195-21f54cfaaf18-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a658cfcc-7de1-467d-8195-21f54cfaaf18" (UID: "a658cfcc-7de1-467d-8195-21f54cfaaf18"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.964485 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a658cfcc-7de1-467d-8195-21f54cfaaf18" (UID: "a658cfcc-7de1-467d-8195-21f54cfaaf18"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.973540 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a658cfcc-7de1-467d-8195-21f54cfaaf18-kube-api-access-hbnnh" (OuterVolumeSpecName: "kube-api-access-hbnnh") pod "a658cfcc-7de1-467d-8195-21f54cfaaf18" (UID: "a658cfcc-7de1-467d-8195-21f54cfaaf18"). InnerVolumeSpecName "kube-api-access-hbnnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:22:57 crc kubenswrapper[4676]: I0124 00:22:57.993558 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-scripts" (OuterVolumeSpecName: "scripts") pod "a658cfcc-7de1-467d-8195-21f54cfaaf18" (UID: "a658cfcc-7de1-467d-8195-21f54cfaaf18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.009265 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a658cfcc-7de1-467d-8195-21f54cfaaf18" (UID: "a658cfcc-7de1-467d-8195-21f54cfaaf18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.018493 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data" (OuterVolumeSpecName: "config-data") pod "a658cfcc-7de1-467d-8195-21f54cfaaf18" (UID: "a658cfcc-7de1-467d-8195-21f54cfaaf18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.058869 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.058909 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.058922 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbnnh\" (UniqueName: \"kubernetes.io/projected/a658cfcc-7de1-467d-8195-21f54cfaaf18-kube-api-access-hbnnh\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.058934 4676 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.058944 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a658cfcc-7de1-467d-8195-21f54cfaaf18-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.058954 4676 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a658cfcc-7de1-467d-8195-21f54cfaaf18-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.201327 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bd579cfd9-q4npp"] Jan 24 00:22:58 crc kubenswrapper[4676]: W0124 00:22:58.203624 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dda8455_6777_4efc_abf3_df547cf58339.slice/crio-4a8cd5c99d2b112e87c52a3a9734d394370cd142a57740512c9257e387d06dab WatchSource:0}: Error finding container 4a8cd5c99d2b112e87c52a3a9734d394370cd142a57740512c9257e387d06dab: Status 404 returned error can't find the container with id 4a8cd5c99d2b112e87c52a3a9734d394370cd142a57740512c9257e387d06dab Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.216324 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-54b7b855f4-s49zw" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.292286 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d7c68bc86-pdds6"] Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.293668 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d7c68bc86-pdds6" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api-log" containerID="cri-o://f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f" gracePeriod=30 Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.293810 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5d7c68bc86-pdds6" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api" containerID="cri-o://dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92" gracePeriod=30 Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.475308 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd579cfd9-q4npp" event={"ID":"6dda8455-6777-4efc-abf3-df547cf58339","Type":"ContainerStarted","Data":"4a8cd5c99d2b112e87c52a3a9734d394370cd142a57740512c9257e387d06dab"} Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.495648 4676 generic.go:334] "Generic (PLEG): container finished" podID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerID="f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f" exitCode=143 Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.495971 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7c68bc86-pdds6" event={"ID":"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4","Type":"ContainerDied","Data":"f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f"} Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.524268 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.524479 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"a658cfcc-7de1-467d-8195-21f54cfaaf18","Type":"ContainerDied","Data":"5f5f73538e6873cbcb22a131253b622d3c4660ae4fc5a284910827195368977f"} Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.524618 4676 scope.go:117] "RemoveContainer" containerID="58af4da8742433d2dc7484bcefd21745ff4c7fa2b61dd06f0a0fc41c705c834a" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.608293 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.620595 4676 scope.go:117] "RemoveContainer" containerID="102eb5e37f96d7d93fb2c20897bc25ad810e3aed4efdc7876fc59651cbc8312e" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.631997 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.671911 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 24 00:22:58 crc kubenswrapper[4676]: E0124 00:22:58.672368 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerName="cinder-api-log" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.672434 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerName="cinder-api-log" Jan 24 00:22:58 crc kubenswrapper[4676]: E0124 00:22:58.672446 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerName="cinder-api" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.672453 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerName="cinder-api" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.672707 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerName="cinder-api" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.672746 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="a658cfcc-7de1-467d-8195-21f54cfaaf18" containerName="cinder-api-log" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.674554 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.680653 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.680799 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.688462 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.688970 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.780734 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.781204 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.781291 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-scripts\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.781371 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38cef52-387b-4633-be75-6dc455ad53c4-logs\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.781518 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d38cef52-387b-4633-be75-6dc455ad53c4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.781600 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.781710 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-config-data-custom\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.781790 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-config-data\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.781868 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv9kc\" (UniqueName: \"kubernetes.io/projected/d38cef52-387b-4633-be75-6dc455ad53c4-kube-api-access-bv9kc\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.883494 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.883725 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-scripts\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.883901 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38cef52-387b-4633-be75-6dc455ad53c4-logs\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.883971 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d38cef52-387b-4633-be75-6dc455ad53c4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.884027 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.884148 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d38cef52-387b-4633-be75-6dc455ad53c4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.884234 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-config-data-custom\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.884269 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-config-data\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.884313 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d38cef52-387b-4633-be75-6dc455ad53c4-logs\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.884320 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv9kc\" (UniqueName: \"kubernetes.io/projected/d38cef52-387b-4633-be75-6dc455ad53c4-kube-api-access-bv9kc\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.884366 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.890988 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-config-data-custom\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.891019 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-scripts\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.891680 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-config-data\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.892036 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.892731 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.909728 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d38cef52-387b-4633-be75-6dc455ad53c4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.910692 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv9kc\" (UniqueName: \"kubernetes.io/projected/d38cef52-387b-4633-be75-6dc455ad53c4-kube-api-access-bv9kc\") pod \"cinder-api-0\" (UID: \"d38cef52-387b-4633-be75-6dc455ad53c4\") " pod="openstack/cinder-api-0" Jan 24 00:22:58 crc kubenswrapper[4676]: I0124 00:22:58.964764 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-b7b989687-7nkv4" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.155:9696/\": dial tcp 10.217.0.155:9696: connect: connection refused" Jan 24 00:22:59 crc kubenswrapper[4676]: I0124 00:22:59.050037 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 24 00:22:59 crc kubenswrapper[4676]: I0124 00:22:59.538990 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd579cfd9-q4npp" event={"ID":"6dda8455-6777-4efc-abf3-df547cf58339","Type":"ContainerStarted","Data":"67be95b90cf93b9933fb1ae2105ea1f08de29440ba90feaccd769f190d427798"} Jan 24 00:22:59 crc kubenswrapper[4676]: I0124 00:22:59.539315 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bd579cfd9-q4npp" event={"ID":"6dda8455-6777-4efc-abf3-df547cf58339","Type":"ContainerStarted","Data":"049dc5462fbdbc52cc39195d99e7a055dad132ed0888c1e53388800217cd3f2b"} Jan 24 00:22:59 crc kubenswrapper[4676]: I0124 00:22:59.540503 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:22:59 crc kubenswrapper[4676]: I0124 00:22:59.572741 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-bd579cfd9-q4npp" podStartSLOduration=3.572727321 podStartE2EDuration="3.572727321s" podCreationTimestamp="2026-01-24 00:22:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:22:59.566566303 +0000 UTC m=+1163.596537304" watchObservedRunningTime="2026-01-24 00:22:59.572727321 +0000 UTC m=+1163.602698322" Jan 24 00:22:59 crc kubenswrapper[4676]: I0124 00:22:59.735675 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.172006 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.172128 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.172700 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"9b6a2140f368edfea407714ce9a9afb7bd7057997de5cfab1861b7a16523894e"} pod="openstack/horizon-bf988b4bd-ls7hp" containerMessage="Container horizon failed startup probe, will be restarted" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.172733 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" containerID="cri-o://9b6a2140f368edfea407714ce9a9afb7bd7057997de5cfab1861b7a16523894e" gracePeriod=30 Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.271492 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a658cfcc-7de1-467d-8195-21f54cfaaf18" path="/var/lib/kubelet/pods/a658cfcc-7de1-467d-8195-21f54cfaaf18/volumes" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.322532 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f876ddf46-fs7qv" podUID="ac7dce6b-3bd9-4ad9-9485-83d9384b8bad" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.322609 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.323321 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"c8c367e4ea3e593fa82e51f67b7f8371ac3844af82364461f69e4513b638a905"} pod="openstack/horizon-f876ddf46-fs7qv" containerMessage="Container horizon failed startup probe, will be restarted" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.323352 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f876ddf46-fs7qv" podUID="ac7dce6b-3bd9-4ad9-9485-83d9384b8bad" containerName="horizon" containerID="cri-o://c8c367e4ea3e593fa82e51f67b7f8371ac3844af82364461f69e4513b638a905" gracePeriod=30 Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.525948 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.556634 4676 generic.go:334] "Generic (PLEG): container finished" podID="2e3ec968-d892-487a-930a-44c79123d54b" containerID="bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724" exitCode=0 Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.556704 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b7b989687-7nkv4" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.556737 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b7b989687-7nkv4" event={"ID":"2e3ec968-d892-487a-930a-44c79123d54b","Type":"ContainerDied","Data":"bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724"} Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.556787 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b7b989687-7nkv4" event={"ID":"2e3ec968-d892-487a-930a-44c79123d54b","Type":"ContainerDied","Data":"59d763231e986383c1ba6301030139cf376841912045d87d7f5c53b19c30ff2a"} Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.556809 4676 scope.go:117] "RemoveContainer" containerID="b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.569255 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d38cef52-387b-4633-be75-6dc455ad53c4","Type":"ContainerStarted","Data":"abfd93d434887a511032777d398f6330f8b3744fee9ab7b6582f92d698b400f1"} Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.569285 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d38cef52-387b-4633-be75-6dc455ad53c4","Type":"ContainerStarted","Data":"b1b11c7d08d075296911e4da993464454c371f2ba04a09e772927aeafdf183b7"} Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.618601 4676 scope.go:117] "RemoveContainer" containerID="bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.625626 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-config\") pod \"2e3ec968-d892-487a-930a-44c79123d54b\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.625783 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-ovndb-tls-certs\") pod \"2e3ec968-d892-487a-930a-44c79123d54b\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.625808 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm9l4\" (UniqueName: \"kubernetes.io/projected/2e3ec968-d892-487a-930a-44c79123d54b-kube-api-access-nm9l4\") pod \"2e3ec968-d892-487a-930a-44c79123d54b\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.625854 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-internal-tls-certs\") pod \"2e3ec968-d892-487a-930a-44c79123d54b\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.625895 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-httpd-config\") pod \"2e3ec968-d892-487a-930a-44c79123d54b\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.625943 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-public-tls-certs\") pod \"2e3ec968-d892-487a-930a-44c79123d54b\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.626027 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-combined-ca-bundle\") pod \"2e3ec968-d892-487a-930a-44c79123d54b\" (UID: \"2e3ec968-d892-487a-930a-44c79123d54b\") " Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.645122 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "2e3ec968-d892-487a-930a-44c79123d54b" (UID: "2e3ec968-d892-487a-930a-44c79123d54b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.646631 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e3ec968-d892-487a-930a-44c79123d54b-kube-api-access-nm9l4" (OuterVolumeSpecName: "kube-api-access-nm9l4") pod "2e3ec968-d892-487a-930a-44c79123d54b" (UID: "2e3ec968-d892-487a-930a-44c79123d54b"). InnerVolumeSpecName "kube-api-access-nm9l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.671208 4676 scope.go:117] "RemoveContainer" containerID="b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae" Jan 24 00:23:00 crc kubenswrapper[4676]: E0124 00:23:00.672877 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae\": container with ID starting with b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae not found: ID does not exist" containerID="b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.672917 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae"} err="failed to get container status \"b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae\": rpc error: code = NotFound desc = could not find container \"b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae\": container with ID starting with b7aee4889e0ac214cfcb4e7b60ee0df9d729146b0da2baea1b44c7b338f2b7ae not found: ID does not exist" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.672940 4676 scope.go:117] "RemoveContainer" containerID="bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724" Jan 24 00:23:00 crc kubenswrapper[4676]: E0124 00:23:00.673417 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724\": container with ID starting with bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724 not found: ID does not exist" containerID="bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.673450 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724"} err="failed to get container status \"bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724\": rpc error: code = NotFound desc = could not find container \"bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724\": container with ID starting with bb52e795a11686953fafadd660026fb29a73517b463940520ebb5925860e8724 not found: ID does not exist" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.723300 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "2e3ec968-d892-487a-930a-44c79123d54b" (UID: "2e3ec968-d892-487a-930a-44c79123d54b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.728904 4676 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.728929 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm9l4\" (UniqueName: \"kubernetes.io/projected/2e3ec968-d892-487a-930a-44c79123d54b-kube-api-access-nm9l4\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.728943 4676 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.729038 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e3ec968-d892-487a-930a-44c79123d54b" (UID: "2e3ec968-d892-487a-930a-44c79123d54b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.734742 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2e3ec968-d892-487a-930a-44c79123d54b" (UID: "2e3ec968-d892-487a-930a-44c79123d54b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.750754 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2e3ec968-d892-487a-930a-44c79123d54b" (UID: "2e3ec968-d892-487a-930a-44c79123d54b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.773477 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-config" (OuterVolumeSpecName: "config") pod "2e3ec968-d892-487a-930a-44c79123d54b" (UID: "2e3ec968-d892-487a-930a-44c79123d54b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.830899 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.830928 4676 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.830940 4676 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.830949 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e3ec968-d892-487a-930a-44c79123d54b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.892509 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b7b989687-7nkv4"] Jan 24 00:23:00 crc kubenswrapper[4676]: I0124 00:23:00.899954 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b7b989687-7nkv4"] Jan 24 00:23:01 crc kubenswrapper[4676]: I0124 00:23:01.579941 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d38cef52-387b-4633-be75-6dc455ad53c4","Type":"ContainerStarted","Data":"c6e0b44a4a2847f30f75aa6d61b861aa105c8c4d85e6a59521835b3f3e845693"} Jan 24 00:23:01 crc kubenswrapper[4676]: I0124 00:23:01.605314 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.605296446 podStartE2EDuration="3.605296446s" podCreationTimestamp="2026-01-24 00:22:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:01.603633025 +0000 UTC m=+1165.633604026" watchObservedRunningTime="2026-01-24 00:23:01.605296446 +0000 UTC m=+1165.635267457" Jan 24 00:23:01 crc kubenswrapper[4676]: I0124 00:23:01.923519 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:23:01 crc kubenswrapper[4676]: I0124 00:23:01.991957 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xg9nt"] Jan 24 00:23:01 crc kubenswrapper[4676]: I0124 00:23:01.992573 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" podUID="e1f791cc-7690-4f80-9007-a14fcca8a632" containerName="dnsmasq-dns" containerID="cri-o://e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b" gracePeriod=10 Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.206359 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.227338 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.267585 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e3ec968-d892-487a-930a-44c79123d54b" path="/var/lib/kubelet/pods/2e3ec968-d892-487a-930a-44c79123d54b/volumes" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.283347 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-combined-ca-bundle\") pod \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.283474 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-logs\") pod \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.283497 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data-custom\") pod \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.283536 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data\") pod \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.283620 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfjfx\" (UniqueName: \"kubernetes.io/projected/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-kube-api-access-bfjfx\") pod \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\" (UID: \"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.284204 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-logs" (OuterVolumeSpecName: "logs") pod "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" (UID: "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.285194 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.292918 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" (UID: "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.300057 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-kube-api-access-bfjfx" (OuterVolumeSpecName: "kube-api-access-bfjfx") pod "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" (UID: "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4"). InnerVolumeSpecName "kube-api-access-bfjfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.340228 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" (UID: "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.376797 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data" (OuterVolumeSpecName: "config-data") pod "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" (UID: "6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.387415 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.387572 4676 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.387620 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.387630 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfjfx\" (UniqueName: \"kubernetes.io/projected/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-kube-api-access-bfjfx\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.387640 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.573632 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.591042 4676 generic.go:334] "Generic (PLEG): container finished" podID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerID="dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92" exitCode=0 Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.591100 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7c68bc86-pdds6" event={"ID":"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4","Type":"ContainerDied","Data":"dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92"} Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.591126 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d7c68bc86-pdds6" event={"ID":"6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4","Type":"ContainerDied","Data":"779fb9c258d57c3a4a4641f3663256b4984fdbbbea094e97da3fbf8189a0afcc"} Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.591143 4676 scope.go:117] "RemoveContainer" containerID="dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.591248 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d7c68bc86-pdds6" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.613454 4676 generic.go:334] "Generic (PLEG): container finished" podID="e1f791cc-7690-4f80-9007-a14fcca8a632" containerID="e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b" exitCode=0 Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.613537 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" event={"ID":"e1f791cc-7690-4f80-9007-a14fcca8a632","Type":"ContainerDied","Data":"e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b"} Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.613602 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.613609 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" event={"ID":"e1f791cc-7690-4f80-9007-a14fcca8a632","Type":"ContainerDied","Data":"2612852ab10c2bc7fb166122d80d7317364b784205eafc69c7981f567d09ca8e"} Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.614153 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.614273 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fb66b31d-c165-4930-b633-79707cbf22bb" containerName="cinder-scheduler" containerID="cri-o://617a7bf60c86bd8fb95135f8ae3c47f3775cca4441a612fa0eecdd996ac43e81" gracePeriod=30 Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.614422 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="fb66b31d-c165-4930-b633-79707cbf22bb" containerName="probe" containerID="cri-o://2401fa1f3b792936f62c0e8fef2c85d82ca1c01d1b0b3520974286bb7dbc9b75" gracePeriod=30 Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.652496 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5d7c68bc86-pdds6"] Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.658431 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5d7c68bc86-pdds6"] Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.661399 4676 scope.go:117] "RemoveContainer" containerID="f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.685589 4676 scope.go:117] "RemoveContainer" containerID="dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92" Jan 24 00:23:02 crc kubenswrapper[4676]: E0124 00:23:02.685952 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92\": container with ID starting with dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92 not found: ID does not exist" containerID="dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.685981 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92"} err="failed to get container status \"dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92\": rpc error: code = NotFound desc = could not find container \"dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92\": container with ID starting with dee85e5f9b90c28514a6fe5264e7105a66d2041ede87770d395b2f20c4c19a92 not found: ID does not exist" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.686004 4676 scope.go:117] "RemoveContainer" containerID="f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f" Jan 24 00:23:02 crc kubenswrapper[4676]: E0124 00:23:02.686190 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f\": container with ID starting with f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f not found: ID does not exist" containerID="f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.686215 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f"} err="failed to get container status \"f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f\": rpc error: code = NotFound desc = could not find container \"f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f\": container with ID starting with f6075d82d7703b6db2b90cc384faadb0429bfb79747fbdd96d4361dda203ec6f not found: ID does not exist" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.686231 4676 scope.go:117] "RemoveContainer" containerID="e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.692566 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-swift-storage-0\") pod \"e1f791cc-7690-4f80-9007-a14fcca8a632\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.692615 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-nb\") pod \"e1f791cc-7690-4f80-9007-a14fcca8a632\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.692794 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-sb\") pod \"e1f791cc-7690-4f80-9007-a14fcca8a632\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.692844 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9j5s9\" (UniqueName: \"kubernetes.io/projected/e1f791cc-7690-4f80-9007-a14fcca8a632-kube-api-access-9j5s9\") pod \"e1f791cc-7690-4f80-9007-a14fcca8a632\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.692904 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-config\") pod \"e1f791cc-7690-4f80-9007-a14fcca8a632\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.692966 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-svc\") pod \"e1f791cc-7690-4f80-9007-a14fcca8a632\" (UID: \"e1f791cc-7690-4f80-9007-a14fcca8a632\") " Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.705339 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1f791cc-7690-4f80-9007-a14fcca8a632-kube-api-access-9j5s9" (OuterVolumeSpecName: "kube-api-access-9j5s9") pod "e1f791cc-7690-4f80-9007-a14fcca8a632" (UID: "e1f791cc-7690-4f80-9007-a14fcca8a632"). InnerVolumeSpecName "kube-api-access-9j5s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.723587 4676 scope.go:117] "RemoveContainer" containerID="cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.761012 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e1f791cc-7690-4f80-9007-a14fcca8a632" (UID: "e1f791cc-7690-4f80-9007-a14fcca8a632"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.761104 4676 scope.go:117] "RemoveContainer" containerID="e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b" Jan 24 00:23:02 crc kubenswrapper[4676]: E0124 00:23:02.761846 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b\": container with ID starting with e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b not found: ID does not exist" containerID="e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.761898 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b"} err="failed to get container status \"e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b\": rpc error: code = NotFound desc = could not find container \"e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b\": container with ID starting with e2b72fc7b5df059c0e36f49a0b10a80c4bc4c2798a40f7961418e5eff05aa92b not found: ID does not exist" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.761932 4676 scope.go:117] "RemoveContainer" containerID="cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a" Jan 24 00:23:02 crc kubenswrapper[4676]: E0124 00:23:02.762241 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a\": container with ID starting with cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a not found: ID does not exist" containerID="cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.762273 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a"} err="failed to get container status \"cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a\": rpc error: code = NotFound desc = could not find container \"cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a\": container with ID starting with cfab6422e88eeccb3646e027bd8862c4d0fb49c299978a170241c30c578c119a not found: ID does not exist" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.768564 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-config" (OuterVolumeSpecName: "config") pod "e1f791cc-7690-4f80-9007-a14fcca8a632" (UID: "e1f791cc-7690-4f80-9007-a14fcca8a632"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.785747 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e1f791cc-7690-4f80-9007-a14fcca8a632" (UID: "e1f791cc-7690-4f80-9007-a14fcca8a632"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.791575 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e1f791cc-7690-4f80-9007-a14fcca8a632" (UID: "e1f791cc-7690-4f80-9007-a14fcca8a632"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.791857 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e1f791cc-7690-4f80-9007-a14fcca8a632" (UID: "e1f791cc-7690-4f80-9007-a14fcca8a632"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.795279 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.795300 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9j5s9\" (UniqueName: \"kubernetes.io/projected/e1f791cc-7690-4f80-9007-a14fcca8a632-kube-api-access-9j5s9\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.795313 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.795325 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.795333 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.795342 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e1f791cc-7690-4f80-9007-a14fcca8a632-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.833866 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.858390 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-765f6cdf58-5q9v9" Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.959486 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xg9nt"] Jan 24 00:23:02 crc kubenswrapper[4676]: I0124 00:23:02.967501 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xg9nt"] Jan 24 00:23:03 crc kubenswrapper[4676]: I0124 00:23:03.631329 4676 generic.go:334] "Generic (PLEG): container finished" podID="fb66b31d-c165-4930-b633-79707cbf22bb" containerID="2401fa1f3b792936f62c0e8fef2c85d82ca1c01d1b0b3520974286bb7dbc9b75" exitCode=0 Jan 24 00:23:03 crc kubenswrapper[4676]: I0124 00:23:03.631410 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fb66b31d-c165-4930-b633-79707cbf22bb","Type":"ContainerDied","Data":"2401fa1f3b792936f62c0e8fef2c85d82ca1c01d1b0b3520974286bb7dbc9b75"} Jan 24 00:23:04 crc kubenswrapper[4676]: I0124 00:23:04.285194 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" path="/var/lib/kubelet/pods/6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4/volumes" Jan 24 00:23:04 crc kubenswrapper[4676]: I0124 00:23:04.286006 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1f791cc-7690-4f80-9007-a14fcca8a632" path="/var/lib/kubelet/pods/e1f791cc-7690-4f80-9007-a14fcca8a632/volumes" Jan 24 00:23:04 crc kubenswrapper[4676]: I0124 00:23:04.667832 4676 generic.go:334] "Generic (PLEG): container finished" podID="fb66b31d-c165-4930-b633-79707cbf22bb" containerID="617a7bf60c86bd8fb95135f8ae3c47f3775cca4441a612fa0eecdd996ac43e81" exitCode=0 Jan 24 00:23:04 crc kubenswrapper[4676]: I0124 00:23:04.668143 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fb66b31d-c165-4930-b633-79707cbf22bb","Type":"ContainerDied","Data":"617a7bf60c86bd8fb95135f8ae3c47f3775cca4441a612fa0eecdd996ac43e81"} Jan 24 00:23:04 crc kubenswrapper[4676]: I0124 00:23:04.961538 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.036430 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-combined-ca-bundle\") pod \"fb66b31d-c165-4930-b633-79707cbf22bb\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.036836 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vknvh\" (UniqueName: \"kubernetes.io/projected/fb66b31d-c165-4930-b633-79707cbf22bb-kube-api-access-vknvh\") pod \"fb66b31d-c165-4930-b633-79707cbf22bb\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.036856 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb66b31d-c165-4930-b633-79707cbf22bb-etc-machine-id\") pod \"fb66b31d-c165-4930-b633-79707cbf22bb\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.036905 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data-custom\") pod \"fb66b31d-c165-4930-b633-79707cbf22bb\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.036930 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-scripts\") pod \"fb66b31d-c165-4930-b633-79707cbf22bb\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.036952 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data\") pod \"fb66b31d-c165-4930-b633-79707cbf22bb\" (UID: \"fb66b31d-c165-4930-b633-79707cbf22bb\") " Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.037494 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb66b31d-c165-4930-b633-79707cbf22bb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fb66b31d-c165-4930-b633-79707cbf22bb" (UID: "fb66b31d-c165-4930-b633-79707cbf22bb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.054698 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb66b31d-c165-4930-b633-79707cbf22bb-kube-api-access-vknvh" (OuterVolumeSpecName: "kube-api-access-vknvh") pod "fb66b31d-c165-4930-b633-79707cbf22bb" (UID: "fb66b31d-c165-4930-b633-79707cbf22bb"). InnerVolumeSpecName "kube-api-access-vknvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.054896 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fb66b31d-c165-4930-b633-79707cbf22bb" (UID: "fb66b31d-c165-4930-b633-79707cbf22bb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.061594 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-scripts" (OuterVolumeSpecName: "scripts") pod "fb66b31d-c165-4930-b633-79707cbf22bb" (UID: "fb66b31d-c165-4930-b633-79707cbf22bb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.103510 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb66b31d-c165-4930-b633-79707cbf22bb" (UID: "fb66b31d-c165-4930-b633-79707cbf22bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.139156 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.139357 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.139687 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vknvh\" (UniqueName: \"kubernetes.io/projected/fb66b31d-c165-4930-b633-79707cbf22bb-kube-api-access-vknvh\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.140074 4676 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fb66b31d-c165-4930-b633-79707cbf22bb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.140153 4676 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.159529 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data" (OuterVolumeSpecName: "config-data") pod "fb66b31d-c165-4930-b633-79707cbf22bb" (UID: "fb66b31d-c165-4930-b633-79707cbf22bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.241972 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb66b31d-c165-4930-b633-79707cbf22bb-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.676819 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fb66b31d-c165-4930-b633-79707cbf22bb","Type":"ContainerDied","Data":"f56885c2863c57838bc58c1bd3891e3e64ca0124e390aeca0b9564b1dc1517dd"} Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.676870 4676 scope.go:117] "RemoveContainer" containerID="2401fa1f3b792936f62c0e8fef2c85d82ca1c01d1b0b3520974286bb7dbc9b75" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.676980 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.702368 4676 scope.go:117] "RemoveContainer" containerID="617a7bf60c86bd8fb95135f8ae3c47f3775cca4441a612fa0eecdd996ac43e81" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.721627 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.737065 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.751355 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 24 00:23:05 crc kubenswrapper[4676]: E0124 00:23:05.752764 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f791cc-7690-4f80-9007-a14fcca8a632" containerName="init" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.752798 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f791cc-7690-4f80-9007-a14fcca8a632" containerName="init" Jan 24 00:23:05 crc kubenswrapper[4676]: E0124 00:23:05.752814 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api-log" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.752822 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api-log" Jan 24 00:23:05 crc kubenswrapper[4676]: E0124 00:23:05.752830 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb66b31d-c165-4930-b633-79707cbf22bb" containerName="probe" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.752836 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb66b31d-c165-4930-b633-79707cbf22bb" containerName="probe" Jan 24 00:23:05 crc kubenswrapper[4676]: E0124 00:23:05.752851 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f791cc-7690-4f80-9007-a14fcca8a632" containerName="dnsmasq-dns" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.752857 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f791cc-7690-4f80-9007-a14fcca8a632" containerName="dnsmasq-dns" Jan 24 00:23:05 crc kubenswrapper[4676]: E0124 00:23:05.752867 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.752872 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api" Jan 24 00:23:05 crc kubenswrapper[4676]: E0124 00:23:05.752896 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb66b31d-c165-4930-b633-79707cbf22bb" containerName="cinder-scheduler" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.752902 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb66b31d-c165-4930-b633-79707cbf22bb" containerName="cinder-scheduler" Jan 24 00:23:05 crc kubenswrapper[4676]: E0124 00:23:05.752911 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-httpd" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.752916 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-httpd" Jan 24 00:23:05 crc kubenswrapper[4676]: E0124 00:23:05.752930 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-api" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.752936 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-api" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.753084 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-api" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.753097 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api-log" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.753108 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e3ec968-d892-487a-930a-44c79123d54b" containerName="neutron-httpd" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.753117 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb66b31d-c165-4930-b633-79707cbf22bb" containerName="cinder-scheduler" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.753133 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb66b31d-c165-4930-b633-79707cbf22bb" containerName="probe" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.753142 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1f791cc-7690-4f80-9007-a14fcca8a632" containerName="dnsmasq-dns" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.753152 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c41d5e8-6a89-452a-bacc-2d7d25cfd6d4" containerName="barbican-api" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.754219 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.759596 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.770597 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.853440 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-264rh\" (UniqueName: \"kubernetes.io/projected/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-kube-api-access-264rh\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.853796 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-scripts\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.853822 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.853861 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-config-data\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.853890 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.854142 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.956117 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-config-data\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.956426 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.956581 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.956770 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-264rh\" (UniqueName: \"kubernetes.io/projected/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-kube-api-access-264rh\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.956864 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-scripts\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.956955 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.956614 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.962356 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.963468 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.964280 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-config-data\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.966914 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-scripts\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:05 crc kubenswrapper[4676]: I0124 00:23:05.993107 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-264rh\" (UniqueName: \"kubernetes.io/projected/629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978-kube-api-access-264rh\") pod \"cinder-scheduler-0\" (UID: \"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978\") " pod="openstack/cinder-scheduler-0" Jan 24 00:23:06 crc kubenswrapper[4676]: I0124 00:23:06.072305 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 24 00:23:06 crc kubenswrapper[4676]: I0124 00:23:06.270008 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb66b31d-c165-4930-b633-79707cbf22bb" path="/var/lib/kubelet/pods/fb66b31d-c165-4930-b633-79707cbf22bb/volumes" Jan 24 00:23:06 crc kubenswrapper[4676]: I0124 00:23:06.514024 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 24 00:23:06 crc kubenswrapper[4676]: W0124 00:23:06.519570 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod629a8ba3_3e4e_4fc2_86a4_bb07d2fb4978.slice/crio-5edb5f7d4665b4762958b02febb9c178c176623212b544a1cdff5460e3f9da71 WatchSource:0}: Error finding container 5edb5f7d4665b4762958b02febb9c178c176623212b544a1cdff5460e3f9da71: Status 404 returned error can't find the container with id 5edb5f7d4665b4762958b02febb9c178c176623212b544a1cdff5460e3f9da71 Jan 24 00:23:06 crc kubenswrapper[4676]: I0124 00:23:06.701671 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978","Type":"ContainerStarted","Data":"5edb5f7d4665b4762958b02febb9c178c176623212b544a1cdff5460e3f9da71"} Jan 24 00:23:07 crc kubenswrapper[4676]: I0124 00:23:07.490558 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-848cf88cfc-xg9nt" podUID="e1f791cc-7690-4f80-9007-a14fcca8a632" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.160:5353: i/o timeout" Jan 24 00:23:07 crc kubenswrapper[4676]: I0124 00:23:07.522835 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-849b597d57-kw79c" Jan 24 00:23:07 crc kubenswrapper[4676]: I0124 00:23:07.725707 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978","Type":"ContainerStarted","Data":"e3340c74f6f22649c021c7e0f187ac176c4b3ae05fe7d091105168fc7ec6f010"} Jan 24 00:23:08 crc kubenswrapper[4676]: I0124 00:23:08.735676 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978","Type":"ContainerStarted","Data":"34dba549c2590def8f9c0cdb1da4e8f0344bb110720c1f6efbec0f56a8694f4b"} Jan 24 00:23:08 crc kubenswrapper[4676]: I0124 00:23:08.772638 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.772620363 podStartE2EDuration="3.772620363s" podCreationTimestamp="2026-01-24 00:23:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:08.762882187 +0000 UTC m=+1172.792853188" watchObservedRunningTime="2026-01-24 00:23:08.772620363 +0000 UTC m=+1172.802591354" Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.364225 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.364279 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.893486 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.895215 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.898229 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.898397 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-swvxc" Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.906578 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.909834 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.938155 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f3fa6e59-785d-4d40-8d73-170552068e43-openstack-config-secret\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.938413 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw4vg\" (UniqueName: \"kubernetes.io/projected/f3fa6e59-785d-4d40-8d73-170552068e43-kube-api-access-nw4vg\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.938545 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fa6e59-785d-4d40-8d73-170552068e43-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:09 crc kubenswrapper[4676]: I0124 00:23:09.938664 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f3fa6e59-785d-4d40-8d73-170552068e43-openstack-config\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.040268 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fa6e59-785d-4d40-8d73-170552068e43-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.040358 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f3fa6e59-785d-4d40-8d73-170552068e43-openstack-config\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.040559 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f3fa6e59-785d-4d40-8d73-170552068e43-openstack-config-secret\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.040603 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw4vg\" (UniqueName: \"kubernetes.io/projected/f3fa6e59-785d-4d40-8d73-170552068e43-kube-api-access-nw4vg\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.041368 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f3fa6e59-785d-4d40-8d73-170552068e43-openstack-config\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.046669 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fa6e59-785d-4d40-8d73-170552068e43-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.060209 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f3fa6e59-785d-4d40-8d73-170552068e43-openstack-config-secret\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.060840 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw4vg\" (UniqueName: \"kubernetes.io/projected/f3fa6e59-785d-4d40-8d73-170552068e43-kube-api-access-nw4vg\") pod \"openstackclient\" (UID: \"f3fa6e59-785d-4d40-8d73-170552068e43\") " pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.224850 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.740841 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 24 00:23:10 crc kubenswrapper[4676]: I0124 00:23:10.768262 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f3fa6e59-785d-4d40-8d73-170552068e43","Type":"ContainerStarted","Data":"3c334f3f512b309d59d7969a994238af066cd6ca2cd2b166ae3f52a6517d0c1f"} Jan 24 00:23:11 crc kubenswrapper[4676]: I0124 00:23:11.072628 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 24 00:23:11 crc kubenswrapper[4676]: I0124 00:23:11.222911 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.075076 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-f47776b4c-v4xb2"] Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.076698 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.077997 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.079938 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.080103 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.090384 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-f47776b4c-v4xb2"] Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.117577 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-config-data\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.117632 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b976b9e2-b80e-4626-919d-3bb84f0151e8-etc-swift\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.117685 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b976b9e2-b80e-4626-919d-3bb84f0151e8-run-httpd\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.117729 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-public-tls-certs\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.117769 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b976b9e2-b80e-4626-919d-3bb84f0151e8-log-httpd\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.117790 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-combined-ca-bundle\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.117835 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-internal-tls-certs\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.117855 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5j5t\" (UniqueName: \"kubernetes.io/projected/b976b9e2-b80e-4626-919d-3bb84f0151e8-kube-api-access-f5j5t\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.219432 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b976b9e2-b80e-4626-919d-3bb84f0151e8-log-httpd\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.219739 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-combined-ca-bundle\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.219802 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-internal-tls-certs\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.219827 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5j5t\" (UniqueName: \"kubernetes.io/projected/b976b9e2-b80e-4626-919d-3bb84f0151e8-kube-api-access-f5j5t\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.219846 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-config-data\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.219873 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b976b9e2-b80e-4626-919d-3bb84f0151e8-etc-swift\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.219919 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b976b9e2-b80e-4626-919d-3bb84f0151e8-run-httpd\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.219961 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-public-tls-certs\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.220626 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b976b9e2-b80e-4626-919d-3bb84f0151e8-log-httpd\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.220945 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b976b9e2-b80e-4626-919d-3bb84f0151e8-run-httpd\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.225910 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-config-data\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.232942 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-internal-tls-certs\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.251997 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-public-tls-certs\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.252126 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5j5t\" (UniqueName: \"kubernetes.io/projected/b976b9e2-b80e-4626-919d-3bb84f0151e8-kube-api-access-f5j5t\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.252892 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b976b9e2-b80e-4626-919d-3bb84f0151e8-etc-swift\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.305904 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b976b9e2-b80e-4626-919d-3bb84f0151e8-combined-ca-bundle\") pod \"swift-proxy-f47776b4c-v4xb2\" (UID: \"b976b9e2-b80e-4626-919d-3bb84f0151e8\") " pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.398804 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:14 crc kubenswrapper[4676]: I0124 00:23:14.985519 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-f47776b4c-v4xb2"] Jan 24 00:23:14 crc kubenswrapper[4676]: W0124 00:23:14.996475 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb976b9e2_b80e_4626_919d_3bb84f0151e8.slice/crio-f33df89d7230ab46e614c7fd4d27c529da899033af6df22183f7274a665d8bfc WatchSource:0}: Error finding container f33df89d7230ab46e614c7fd4d27c529da899033af6df22183f7274a665d8bfc: Status 404 returned error can't find the container with id f33df89d7230ab46e614c7fd4d27c529da899033af6df22183f7274a665d8bfc Jan 24 00:23:15 crc kubenswrapper[4676]: I0124 00:23:15.820341 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-f47776b4c-v4xb2" event={"ID":"b976b9e2-b80e-4626-919d-3bb84f0151e8","Type":"ContainerStarted","Data":"998a1d998a7ec6be502f68d6f41bc73bc392c7fc5d51b74f3156c1e9d3e06f50"} Jan 24 00:23:15 crc kubenswrapper[4676]: I0124 00:23:15.820916 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-f47776b4c-v4xb2" event={"ID":"b976b9e2-b80e-4626-919d-3bb84f0151e8","Type":"ContainerStarted","Data":"00d5f9d23144a1151738a7c3d52da137802c935773f92f51e8a47f56cb4e8894"} Jan 24 00:23:15 crc kubenswrapper[4676]: I0124 00:23:15.820927 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-f47776b4c-v4xb2" event={"ID":"b976b9e2-b80e-4626-919d-3bb84f0151e8","Type":"ContainerStarted","Data":"f33df89d7230ab46e614c7fd4d27c529da899033af6df22183f7274a665d8bfc"} Jan 24 00:23:16 crc kubenswrapper[4676]: I0124 00:23:16.333981 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 24 00:23:16 crc kubenswrapper[4676]: I0124 00:23:16.826888 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:16 crc kubenswrapper[4676]: I0124 00:23:16.826968 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:16 crc kubenswrapper[4676]: I0124 00:23:16.843798 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-f47776b4c-v4xb2" podStartSLOduration=2.8437788680000002 podStartE2EDuration="2.843778868s" podCreationTimestamp="2026-01-24 00:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:16.84186989 +0000 UTC m=+1180.871840891" watchObservedRunningTime="2026-01-24 00:23:16.843778868 +0000 UTC m=+1180.873749869" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.002854 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-bpr95"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.005009 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.018052 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bpr95"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.147040 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wl59\" (UniqueName: \"kubernetes.io/projected/ee552380-8d96-4a10-a5b8-2deb2e73b15f-kube-api-access-2wl59\") pod \"nova-api-db-create-bpr95\" (UID: \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\") " pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.147133 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee552380-8d96-4a10-a5b8-2deb2e73b15f-operator-scripts\") pod \"nova-api-db-create-bpr95\" (UID: \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\") " pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.166811 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-vtxzm"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.167889 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.190348 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vtxzm"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.248393 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee552380-8d96-4a10-a5b8-2deb2e73b15f-operator-scripts\") pod \"nova-api-db-create-bpr95\" (UID: \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\") " pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.248507 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wl59\" (UniqueName: \"kubernetes.io/projected/ee552380-8d96-4a10-a5b8-2deb2e73b15f-kube-api-access-2wl59\") pod \"nova-api-db-create-bpr95\" (UID: \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\") " pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.249467 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee552380-8d96-4a10-a5b8-2deb2e73b15f-operator-scripts\") pod \"nova-api-db-create-bpr95\" (UID: \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\") " pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.293602 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-a963-account-create-update-jx4g6"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.294687 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.297714 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a963-account-create-update-jx4g6"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.300709 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.304202 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wl59\" (UniqueName: \"kubernetes.io/projected/ee552380-8d96-4a10-a5b8-2deb2e73b15f-kube-api-access-2wl59\") pod \"nova-api-db-create-bpr95\" (UID: \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\") " pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.328020 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.334419 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-b7wzs"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.335576 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.356025 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-727rf\" (UniqueName: \"kubernetes.io/projected/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-kube-api-access-727rf\") pod \"nova-cell0-db-create-vtxzm\" (UID: \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\") " pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.356160 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-operator-scripts\") pod \"nova-cell0-db-create-vtxzm\" (UID: \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\") " pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.374747 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b7wzs"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.459050 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-operator-scripts\") pod \"nova-cell0-db-create-vtxzm\" (UID: \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\") " pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.471782 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nljvc\" (UniqueName: \"kubernetes.io/projected/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-kube-api-access-nljvc\") pod \"nova-api-a963-account-create-update-jx4g6\" (UID: \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\") " pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.473194 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-operator-scripts\") pod \"nova-cell1-db-create-b7wzs\" (UID: \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\") " pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.473337 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfkb7\" (UniqueName: \"kubernetes.io/projected/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-kube-api-access-cfkb7\") pod \"nova-cell1-db-create-b7wzs\" (UID: \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\") " pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.473451 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-operator-scripts\") pod \"nova-api-a963-account-create-update-jx4g6\" (UID: \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\") " pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.473590 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-727rf\" (UniqueName: \"kubernetes.io/projected/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-kube-api-access-727rf\") pod \"nova-cell0-db-create-vtxzm\" (UID: \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\") " pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.463786 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-operator-scripts\") pod \"nova-cell0-db-create-vtxzm\" (UID: \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\") " pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.519023 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-727rf\" (UniqueName: \"kubernetes.io/projected/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-kube-api-access-727rf\") pod \"nova-cell0-db-create-vtxzm\" (UID: \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\") " pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.578209 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-operator-scripts\") pod \"nova-api-a963-account-create-update-jx4g6\" (UID: \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\") " pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.578345 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nljvc\" (UniqueName: \"kubernetes.io/projected/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-kube-api-access-nljvc\") pod \"nova-api-a963-account-create-update-jx4g6\" (UID: \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\") " pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.578396 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-operator-scripts\") pod \"nova-cell1-db-create-b7wzs\" (UID: \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\") " pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.578432 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfkb7\" (UniqueName: \"kubernetes.io/projected/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-kube-api-access-cfkb7\") pod \"nova-cell1-db-create-b7wzs\" (UID: \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\") " pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.579243 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-operator-scripts\") pod \"nova-api-a963-account-create-update-jx4g6\" (UID: \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\") " pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.579849 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-operator-scripts\") pod \"nova-cell1-db-create-b7wzs\" (UID: \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\") " pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.624126 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-95bf-account-create-update-7cvdb"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.625135 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.625889 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nljvc\" (UniqueName: \"kubernetes.io/projected/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-kube-api-access-nljvc\") pod \"nova-api-a963-account-create-update-jx4g6\" (UID: \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\") " pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.631081 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.638100 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfkb7\" (UniqueName: \"kubernetes.io/projected/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-kube-api-access-cfkb7\") pod \"nova-cell1-db-create-b7wzs\" (UID: \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\") " pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.662962 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-95bf-account-create-update-7cvdb"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.678056 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.679526 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a5f8ec9-ee62-4078-846c-291a47631ffb-operator-scripts\") pod \"nova-cell0-95bf-account-create-update-7cvdb\" (UID: \"2a5f8ec9-ee62-4078-846c-291a47631ffb\") " pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.679680 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmmfv\" (UniqueName: \"kubernetes.io/projected/2a5f8ec9-ee62-4078-846c-291a47631ffb-kube-api-access-qmmfv\") pod \"nova-cell0-95bf-account-create-update-7cvdb\" (UID: \"2a5f8ec9-ee62-4078-846c-291a47631ffb\") " pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.697563 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.723903 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5f9c-account-create-update-9tmf6"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.727826 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.732935 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.747015 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5f9c-account-create-update-9tmf6"] Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.781074 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2ngb\" (UniqueName: \"kubernetes.io/projected/bd4662bb-5179-4f57-8571-5198dcf69bdb-kube-api-access-g2ngb\") pod \"nova-cell1-5f9c-account-create-update-9tmf6\" (UID: \"bd4662bb-5179-4f57-8571-5198dcf69bdb\") " pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.781199 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmmfv\" (UniqueName: \"kubernetes.io/projected/2a5f8ec9-ee62-4078-846c-291a47631ffb-kube-api-access-qmmfv\") pod \"nova-cell0-95bf-account-create-update-7cvdb\" (UID: \"2a5f8ec9-ee62-4078-846c-291a47631ffb\") " pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.781244 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4662bb-5179-4f57-8571-5198dcf69bdb-operator-scripts\") pod \"nova-cell1-5f9c-account-create-update-9tmf6\" (UID: \"bd4662bb-5179-4f57-8571-5198dcf69bdb\") " pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.781266 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a5f8ec9-ee62-4078-846c-291a47631ffb-operator-scripts\") pod \"nova-cell0-95bf-account-create-update-7cvdb\" (UID: \"2a5f8ec9-ee62-4078-846c-291a47631ffb\") " pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.781887 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a5f8ec9-ee62-4078-846c-291a47631ffb-operator-scripts\") pod \"nova-cell0-95bf-account-create-update-7cvdb\" (UID: \"2a5f8ec9-ee62-4078-846c-291a47631ffb\") " pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.794780 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.812034 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmmfv\" (UniqueName: \"kubernetes.io/projected/2a5f8ec9-ee62-4078-846c-291a47631ffb-kube-api-access-qmmfv\") pod \"nova-cell0-95bf-account-create-update-7cvdb\" (UID: \"2a5f8ec9-ee62-4078-846c-291a47631ffb\") " pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.882538 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4662bb-5179-4f57-8571-5198dcf69bdb-operator-scripts\") pod \"nova-cell1-5f9c-account-create-update-9tmf6\" (UID: \"bd4662bb-5179-4f57-8571-5198dcf69bdb\") " pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.882831 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2ngb\" (UniqueName: \"kubernetes.io/projected/bd4662bb-5179-4f57-8571-5198dcf69bdb-kube-api-access-g2ngb\") pod \"nova-cell1-5f9c-account-create-update-9tmf6\" (UID: \"bd4662bb-5179-4f57-8571-5198dcf69bdb\") " pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.883246 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4662bb-5179-4f57-8571-5198dcf69bdb-operator-scripts\") pod \"nova-cell1-5f9c-account-create-update-9tmf6\" (UID: \"bd4662bb-5179-4f57-8571-5198dcf69bdb\") " pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:20 crc kubenswrapper[4676]: I0124 00:23:20.906930 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2ngb\" (UniqueName: \"kubernetes.io/projected/bd4662bb-5179-4f57-8571-5198dcf69bdb-kube-api-access-g2ngb\") pod \"nova-cell1-5f9c-account-create-update-9tmf6\" (UID: \"bd4662bb-5179-4f57-8571-5198dcf69bdb\") " pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:21 crc kubenswrapper[4676]: I0124 00:23:21.041087 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:21 crc kubenswrapper[4676]: I0124 00:23:21.048500 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:21 crc kubenswrapper[4676]: I0124 00:23:21.906893 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:23:21 crc kubenswrapper[4676]: I0124 00:23:21.907140 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="643e6d41-6572-4f21-8651-7f577967bfe8" containerName="glance-log" containerID="cri-o://b6416b75b9719da0d296633782ac2fc04e1dd20b2e9d142f5025be8c5a5d754d" gracePeriod=30 Jan 24 00:23:21 crc kubenswrapper[4676]: I0124 00:23:21.907274 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="643e6d41-6572-4f21-8651-7f577967bfe8" containerName="glance-httpd" containerID="cri-o://e3769b8edac359e6c5b4b6e492f9b2bb7ee70731b0b4c2778ff42c7ddf474ff8" gracePeriod=30 Jan 24 00:23:22 crc kubenswrapper[4676]: I0124 00:23:22.108224 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 24 00:23:22 crc kubenswrapper[4676]: I0124 00:23:22.854521 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:23:22 crc kubenswrapper[4676]: I0124 00:23:22.854805 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="77ec0563-46f1-45b0-892b-352d088f9517" containerName="glance-httpd" containerID="cri-o://1989fb85f0013ac25e4d8fb0c72cdef0a1a08075037f574a25a6701244b8dd50" gracePeriod=30 Jan 24 00:23:22 crc kubenswrapper[4676]: I0124 00:23:22.854752 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="77ec0563-46f1-45b0-892b-352d088f9517" containerName="glance-log" containerID="cri-o://7c7d07ff8323e9ccead03e4d66a066ad64d46892c577bd47375ca06cbd3b2418" gracePeriod=30 Jan 24 00:23:22 crc kubenswrapper[4676]: I0124 00:23:22.878026 4676 generic.go:334] "Generic (PLEG): container finished" podID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerID="1b26cefde45473bb50d47ba50a2ec78ece32f247247ca63eb469045a70b4a673" exitCode=137 Jan 24 00:23:22 crc kubenswrapper[4676]: I0124 00:23:22.878079 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d56910-d4b7-45b1-8699-5af7eaa29b96","Type":"ContainerDied","Data":"1b26cefde45473bb50d47ba50a2ec78ece32f247247ca63eb469045a70b4a673"} Jan 24 00:23:22 crc kubenswrapper[4676]: I0124 00:23:22.881742 4676 generic.go:334] "Generic (PLEG): container finished" podID="643e6d41-6572-4f21-8651-7f577967bfe8" containerID="b6416b75b9719da0d296633782ac2fc04e1dd20b2e9d142f5025be8c5a5d754d" exitCode=143 Jan 24 00:23:22 crc kubenswrapper[4676]: I0124 00:23:22.881768 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"643e6d41-6572-4f21-8651-7f577967bfe8","Type":"ContainerDied","Data":"b6416b75b9719da0d296633782ac2fc04e1dd20b2e9d142f5025be8c5a5d754d"} Jan 24 00:23:23 crc kubenswrapper[4676]: I0124 00:23:23.896729 4676 generic.go:334] "Generic (PLEG): container finished" podID="77ec0563-46f1-45b0-892b-352d088f9517" containerID="7c7d07ff8323e9ccead03e4d66a066ad64d46892c577bd47375ca06cbd3b2418" exitCode=143 Jan 24 00:23:23 crc kubenswrapper[4676]: I0124 00:23:23.896771 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"77ec0563-46f1-45b0-892b-352d088f9517","Type":"ContainerDied","Data":"7c7d07ff8323e9ccead03e4d66a066ad64d46892c577bd47375ca06cbd3b2418"} Jan 24 00:23:24 crc kubenswrapper[4676]: I0124 00:23:24.415562 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:24 crc kubenswrapper[4676]: I0124 00:23:24.423840 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-f47776b4c-v4xb2" Jan 24 00:23:24 crc kubenswrapper[4676]: I0124 00:23:24.717971 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-95bf-account-create-update-7cvdb"] Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.207733 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.387291 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-combined-ca-bundle\") pod \"54d56910-d4b7-45b1-8699-5af7eaa29b96\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.387418 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tljs7\" (UniqueName: \"kubernetes.io/projected/54d56910-d4b7-45b1-8699-5af7eaa29b96-kube-api-access-tljs7\") pod \"54d56910-d4b7-45b1-8699-5af7eaa29b96\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.387457 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-config-data\") pod \"54d56910-d4b7-45b1-8699-5af7eaa29b96\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.387525 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-sg-core-conf-yaml\") pod \"54d56910-d4b7-45b1-8699-5af7eaa29b96\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.387585 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-scripts\") pod \"54d56910-d4b7-45b1-8699-5af7eaa29b96\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.387614 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-log-httpd\") pod \"54d56910-d4b7-45b1-8699-5af7eaa29b96\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.387662 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-run-httpd\") pod \"54d56910-d4b7-45b1-8699-5af7eaa29b96\" (UID: \"54d56910-d4b7-45b1-8699-5af7eaa29b96\") " Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.394368 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "54d56910-d4b7-45b1-8699-5af7eaa29b96" (UID: "54d56910-d4b7-45b1-8699-5af7eaa29b96"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.395775 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "54d56910-d4b7-45b1-8699-5af7eaa29b96" (UID: "54d56910-d4b7-45b1-8699-5af7eaa29b96"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.396771 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54d56910-d4b7-45b1-8699-5af7eaa29b96-kube-api-access-tljs7" (OuterVolumeSpecName: "kube-api-access-tljs7") pod "54d56910-d4b7-45b1-8699-5af7eaa29b96" (UID: "54d56910-d4b7-45b1-8699-5af7eaa29b96"). InnerVolumeSpecName "kube-api-access-tljs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.402518 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-scripts" (OuterVolumeSpecName: "scripts") pod "54d56910-d4b7-45b1-8699-5af7eaa29b96" (UID: "54d56910-d4b7-45b1-8699-5af7eaa29b96"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.489067 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.489091 4676 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.489099 4676 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/54d56910-d4b7-45b1-8699-5af7eaa29b96-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.489108 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tljs7\" (UniqueName: \"kubernetes.io/projected/54d56910-d4b7-45b1-8699-5af7eaa29b96-kube-api-access-tljs7\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:25 crc kubenswrapper[4676]: E0124 00:23:25.528140 4676 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod643e6d41_6572_4f21_8651_7f577967bfe8.slice/crio-e3769b8edac359e6c5b4b6e492f9b2bb7ee70731b0b4c2778ff42c7ddf474ff8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod643e6d41_6572_4f21_8651_7f577967bfe8.slice/crio-conmon-e3769b8edac359e6c5b4b6e492f9b2bb7ee70731b0b4c2778ff42c7ddf474ff8.scope\": RecentStats: unable to find data in memory cache]" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.532955 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bpr95"] Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.549010 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-vtxzm"] Jan 24 00:23:25 crc kubenswrapper[4676]: W0124 00:23:25.569717 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee552380_8d96_4a10_a5b8_2deb2e73b15f.slice/crio-0b915aba37b3b554522a226fb9001f30eb50da19436816f3096df59862d25de5 WatchSource:0}: Error finding container 0b915aba37b3b554522a226fb9001f30eb50da19436816f3096df59862d25de5: Status 404 returned error can't find the container with id 0b915aba37b3b554522a226fb9001f30eb50da19436816f3096df59862d25de5 Jan 24 00:23:25 crc kubenswrapper[4676]: W0124 00:23:25.587926 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod792cc535_5ed4_4e06_a19c_31ba34c7dfc7.slice/crio-82cc1cf40bad027637ba471334a5a9f3e4778b000f0e88aaab517a858e31e686 WatchSource:0}: Error finding container 82cc1cf40bad027637ba471334a5a9f3e4778b000f0e88aaab517a858e31e686: Status 404 returned error can't find the container with id 82cc1cf40bad027637ba471334a5a9f3e4778b000f0e88aaab517a858e31e686 Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.704242 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "54d56910-d4b7-45b1-8699-5af7eaa29b96" (UID: "54d56910-d4b7-45b1-8699-5af7eaa29b96"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.772026 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-b7wzs"] Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.796705 4676 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.843665 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "54d56910-d4b7-45b1-8699-5af7eaa29b96" (UID: "54d56910-d4b7-45b1-8699-5af7eaa29b96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.899408 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.924365 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-config-data" (OuterVolumeSpecName: "config-data") pod "54d56910-d4b7-45b1-8699-5af7eaa29b96" (UID: "54d56910-d4b7-45b1-8699-5af7eaa29b96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.928024 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5f9c-account-create-update-9tmf6"] Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.956566 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" event={"ID":"2a5f8ec9-ee62-4078-846c-291a47631ffb","Type":"ContainerStarted","Data":"27297ea6827aaabd3e76dbca0980a960477f82e37ca6b6bb47dba9c165652738"} Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.956602 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" event={"ID":"2a5f8ec9-ee62-4078-846c-291a47631ffb","Type":"ContainerStarted","Data":"e22d8a5b2f69571d42047489f1547c9fe910b6a190e614d344f927ed2bce6dc6"} Jan 24 00:23:25 crc kubenswrapper[4676]: I0124 00:23:25.977290 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b7wzs" event={"ID":"9cd4cfd3-72ab-45d6-8683-85acb6cadf66","Type":"ContainerStarted","Data":"19a2176ef505fcf59da905e70e48985ee4a9f5afc8429b0b5550316e0527bf2e"} Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.001888 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54d56910-d4b7-45b1-8699-5af7eaa29b96-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.010692 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-a963-account-create-update-jx4g6"] Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.014598 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f3fa6e59-785d-4d40-8d73-170552068e43","Type":"ContainerStarted","Data":"db60a43a3269c8d0b17e12f1f8dac812f95ded60a81f6b6b73124658d3fff20a"} Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.022051 4676 generic.go:334] "Generic (PLEG): container finished" podID="643e6d41-6572-4f21-8651-7f577967bfe8" containerID="e3769b8edac359e6c5b4b6e492f9b2bb7ee70731b0b4c2778ff42c7ddf474ff8" exitCode=0 Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.022112 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"643e6d41-6572-4f21-8651-7f577967bfe8","Type":"ContainerDied","Data":"e3769b8edac359e6c5b4b6e492f9b2bb7ee70731b0b4c2778ff42c7ddf474ff8"} Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.031240 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"54d56910-d4b7-45b1-8699-5af7eaa29b96","Type":"ContainerDied","Data":"3a171779e2acfaaedcae6bf53bf013bc9cc89a659479afefe0a8349e50f12f73"} Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.031282 4676 scope.go:117] "RemoveContainer" containerID="1b26cefde45473bb50d47ba50a2ec78ece32f247247ca63eb469045a70b4a673" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.031409 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.037078 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" podStartSLOduration=6.03705786 podStartE2EDuration="6.03705786s" podCreationTimestamp="2026-01-24 00:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:25.987735099 +0000 UTC m=+1190.017706100" watchObservedRunningTime="2026-01-24 00:23:26.03705786 +0000 UTC m=+1190.067028861" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.054699 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vtxzm" event={"ID":"792cc535-5ed4-4e06-a19c-31ba34c7dfc7","Type":"ContainerStarted","Data":"82cc1cf40bad027637ba471334a5a9f3e4778b000f0e88aaab517a858e31e686"} Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.060873 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bpr95" event={"ID":"ee552380-8d96-4a10-a5b8-2deb2e73b15f","Type":"ContainerStarted","Data":"0b915aba37b3b554522a226fb9001f30eb50da19436816f3096df59862d25de5"} Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.123565 4676 scope.go:117] "RemoveContainer" containerID="dae61f7a8060ccdb486fd53e86c1fca33bbb71e0fb9f0d456ad2e0a5a8007cbd" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.135125 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.935849166 podStartE2EDuration="17.135108642s" podCreationTimestamp="2026-01-24 00:23:09 +0000 UTC" firstStartedPulling="2026-01-24 00:23:10.756110546 +0000 UTC m=+1174.786081547" lastFinishedPulling="2026-01-24 00:23:24.955370022 +0000 UTC m=+1188.985341023" observedRunningTime="2026-01-24 00:23:26.034713169 +0000 UTC m=+1190.064684170" watchObservedRunningTime="2026-01-24 00:23:26.135108642 +0000 UTC m=+1190.165079643" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.365296 4676 scope.go:117] "RemoveContainer" containerID="f53a3c895796c44b94d61775eb726821f040873261a1a8fbdb79f578ca085ec5" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.457124 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.519653 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-scripts\") pod \"643e6d41-6572-4f21-8651-7f577967bfe8\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.519718 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blvls\" (UniqueName: \"kubernetes.io/projected/643e6d41-6572-4f21-8651-7f577967bfe8-kube-api-access-blvls\") pod \"643e6d41-6572-4f21-8651-7f577967bfe8\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.519880 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-httpd-run\") pod \"643e6d41-6572-4f21-8651-7f577967bfe8\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.519934 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-combined-ca-bundle\") pod \"643e6d41-6572-4f21-8651-7f577967bfe8\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.519975 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"643e6d41-6572-4f21-8651-7f577967bfe8\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.520004 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-logs\") pod \"643e6d41-6572-4f21-8651-7f577967bfe8\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.520023 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-public-tls-certs\") pod \"643e6d41-6572-4f21-8651-7f577967bfe8\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.520060 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-config-data\") pod \"643e6d41-6572-4f21-8651-7f577967bfe8\" (UID: \"643e6d41-6572-4f21-8651-7f577967bfe8\") " Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.522042 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-logs" (OuterVolumeSpecName: "logs") pod "643e6d41-6572-4f21-8651-7f577967bfe8" (UID: "643e6d41-6572-4f21-8651-7f577967bfe8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.535433 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "643e6d41-6572-4f21-8651-7f577967bfe8" (UID: "643e6d41-6572-4f21-8651-7f577967bfe8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.545174 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-scripts" (OuterVolumeSpecName: "scripts") pod "643e6d41-6572-4f21-8651-7f577967bfe8" (UID: "643e6d41-6572-4f21-8651-7f577967bfe8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.552977 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "643e6d41-6572-4f21-8651-7f577967bfe8" (UID: "643e6d41-6572-4f21-8651-7f577967bfe8"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.621970 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.622288 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.622297 4676 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/643e6d41-6572-4f21-8651-7f577967bfe8-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.622325 4676 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.667037 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/643e6d41-6572-4f21-8651-7f577967bfe8-kube-api-access-blvls" (OuterVolumeSpecName: "kube-api-access-blvls") pod "643e6d41-6572-4f21-8651-7f577967bfe8" (UID: "643e6d41-6572-4f21-8651-7f577967bfe8"). InnerVolumeSpecName "kube-api-access-blvls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.723636 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blvls\" (UniqueName: \"kubernetes.io/projected/643e6d41-6572-4f21-8651-7f577967bfe8-kube-api-access-blvls\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.758970 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "643e6d41-6572-4f21-8651-7f577967bfe8" (UID: "643e6d41-6572-4f21-8651-7f577967bfe8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.766702 4676 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.792932 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "643e6d41-6572-4f21-8651-7f577967bfe8" (UID: "643e6d41-6572-4f21-8651-7f577967bfe8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.835267 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.835291 4676 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.835299 4676 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:26 crc kubenswrapper[4676]: I0124 00:23:26.998382 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-config-data" (OuterVolumeSpecName: "config-data") pod "643e6d41-6572-4f21-8651-7f577967bfe8" (UID: "643e6d41-6572-4f21-8651-7f577967bfe8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.038957 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/643e6d41-6572-4f21-8651-7f577967bfe8-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.081860 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a963-account-create-update-jx4g6" event={"ID":"7d47faf0-8eaa-474f-8cde-aed6c10f4a05","Type":"ContainerStarted","Data":"cc3b8c014c9450ef0997c6ac2396003671e4bb4fb9c8343ce0debdf9d3796b15"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.081929 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a963-account-create-update-jx4g6" event={"ID":"7d47faf0-8eaa-474f-8cde-aed6c10f4a05","Type":"ContainerStarted","Data":"80f0130f0adbb814673f5191c8d75bf65643388a6bb0829cd96603851a909904"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.083341 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b7wzs" event={"ID":"9cd4cfd3-72ab-45d6-8683-85acb6cadf66","Type":"ContainerStarted","Data":"848f0e8f0de2d0ecec1629bf16e2fb50f2be26fc04335b7c112ae212cbf9035a"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.104598 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"643e6d41-6572-4f21-8651-7f577967bfe8","Type":"ContainerDied","Data":"7eaab757b11ce1880975cf769167065cfd261eeb0ede6ebf82f2d90ebe083765"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.104650 4676 scope.go:117] "RemoveContainer" containerID="e3769b8edac359e6c5b4b6e492f9b2bb7ee70731b0b4c2778ff42c7ddf474ff8" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.104615 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.119220 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" event={"ID":"bd4662bb-5179-4f57-8571-5198dcf69bdb","Type":"ContainerStarted","Data":"0b76bccfcacd7fb1c879fa9b8c0e01b681d270afe951e54be9117c071c170135"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.119270 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" event={"ID":"bd4662bb-5179-4f57-8571-5198dcf69bdb","Type":"ContainerStarted","Data":"e0549f129beec80d104c722afe731613b69d0b4983a48a2b5cd879221ad0c238"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.136235 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vtxzm" event={"ID":"792cc535-5ed4-4e06-a19c-31ba34c7dfc7","Type":"ContainerStarted","Data":"3ee83666354e8c207624c3af1d827bd9c61811a7455e005083ff4e2e34ef0d81"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.159721 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bpr95" event={"ID":"ee552380-8d96-4a10-a5b8-2deb2e73b15f","Type":"ContainerStarted","Data":"e7fc18fa4adecb5d9d62a155b3da665c4c438818bc9b1324ec05947c45e02af0"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.160528 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-b7wzs" podStartSLOduration=7.160512206 podStartE2EDuration="7.160512206s" podCreationTimestamp="2026-01-24 00:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:27.125721279 +0000 UTC m=+1191.155692290" watchObservedRunningTime="2026-01-24 00:23:27.160512206 +0000 UTC m=+1191.190483207" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.161793 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-a963-account-create-update-jx4g6" podStartSLOduration=7.161786185 podStartE2EDuration="7.161786185s" podCreationTimestamp="2026-01-24 00:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:27.103794912 +0000 UTC m=+1191.133765913" watchObservedRunningTime="2026-01-24 00:23:27.161786185 +0000 UTC m=+1191.191757186" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.176998 4676 generic.go:334] "Generic (PLEG): container finished" podID="2a5f8ec9-ee62-4078-846c-291a47631ffb" containerID="27297ea6827aaabd3e76dbca0980a960477f82e37ca6b6bb47dba9c165652738" exitCode=0 Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.177101 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" event={"ID":"2a5f8ec9-ee62-4078-846c-291a47631ffb","Type":"ContainerDied","Data":"27297ea6827aaabd3e76dbca0980a960477f82e37ca6b6bb47dba9c165652738"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.196886 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" podStartSLOduration=7.196867792 podStartE2EDuration="7.196867792s" podCreationTimestamp="2026-01-24 00:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:27.159686442 +0000 UTC m=+1191.189657443" watchObservedRunningTime="2026-01-24 00:23:27.196867792 +0000 UTC m=+1191.226838793" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.199814 4676 generic.go:334] "Generic (PLEG): container finished" podID="77ec0563-46f1-45b0-892b-352d088f9517" containerID="1989fb85f0013ac25e4d8fb0c72cdef0a1a08075037f574a25a6701244b8dd50" exitCode=0 Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.200614 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"77ec0563-46f1-45b0-892b-352d088f9517","Type":"ContainerDied","Data":"1989fb85f0013ac25e4d8fb0c72cdef0a1a08075037f574a25a6701244b8dd50"} Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.206923 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.207855 4676 scope.go:117] "RemoveContainer" containerID="b6416b75b9719da0d296633782ac2fc04e1dd20b2e9d142f5025be8c5a5d754d" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.214488 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-bpr95" podStartSLOduration=8.214471487 podStartE2EDuration="8.214471487s" podCreationTimestamp="2026-01-24 00:23:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:27.190161778 +0000 UTC m=+1191.220132779" watchObservedRunningTime="2026-01-24 00:23:27.214471487 +0000 UTC m=+1191.244442488" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.236750 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-vtxzm" podStartSLOduration=7.236734585 podStartE2EDuration="7.236734585s" podCreationTimestamp="2026-01-24 00:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:27.204887947 +0000 UTC m=+1191.234858948" watchObservedRunningTime="2026-01-24 00:23:27.236734585 +0000 UTC m=+1191.266705576" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.251037 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-internal-tls-certs\") pod \"77ec0563-46f1-45b0-892b-352d088f9517\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.251092 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-httpd-run\") pod \"77ec0563-46f1-45b0-892b-352d088f9517\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.251110 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-config-data\") pod \"77ec0563-46f1-45b0-892b-352d088f9517\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.251127 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"77ec0563-46f1-45b0-892b-352d088f9517\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.251143 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-logs\") pod \"77ec0563-46f1-45b0-892b-352d088f9517\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.251159 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-scripts\") pod \"77ec0563-46f1-45b0-892b-352d088f9517\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.251233 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-combined-ca-bundle\") pod \"77ec0563-46f1-45b0-892b-352d088f9517\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.251256 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49fqw\" (UniqueName: \"kubernetes.io/projected/77ec0563-46f1-45b0-892b-352d088f9517-kube-api-access-49fqw\") pod \"77ec0563-46f1-45b0-892b-352d088f9517\" (UID: \"77ec0563-46f1-45b0-892b-352d088f9517\") " Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.252599 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-logs" (OuterVolumeSpecName: "logs") pod "77ec0563-46f1-45b0-892b-352d088f9517" (UID: "77ec0563-46f1-45b0-892b-352d088f9517"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.253083 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.262497 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "77ec0563-46f1-45b0-892b-352d088f9517" (UID: "77ec0563-46f1-45b0-892b-352d088f9517"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.291982 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-scripts" (OuterVolumeSpecName: "scripts") pod "77ec0563-46f1-45b0-892b-352d088f9517" (UID: "77ec0563-46f1-45b0-892b-352d088f9517"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.292084 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "77ec0563-46f1-45b0-892b-352d088f9517" (UID: "77ec0563-46f1-45b0-892b-352d088f9517"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.324182 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77ec0563-46f1-45b0-892b-352d088f9517-kube-api-access-49fqw" (OuterVolumeSpecName: "kube-api-access-49fqw") pod "77ec0563-46f1-45b0-892b-352d088f9517" (UID: "77ec0563-46f1-45b0-892b-352d088f9517"). InnerVolumeSpecName "kube-api-access-49fqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.354250 4676 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/77ec0563-46f1-45b0-892b-352d088f9517-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.354280 4676 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.354289 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.354298 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49fqw\" (UniqueName: \"kubernetes.io/projected/77ec0563-46f1-45b0-892b-352d088f9517-kube-api-access-49fqw\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.362831 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.363232 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "77ec0563-46f1-45b0-892b-352d088f9517" (UID: "77ec0563-46f1-45b0-892b-352d088f9517"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.373654 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.387674 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-bd579cfd9-q4npp" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.391740 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-config-data" (OuterVolumeSpecName: "config-data") pod "77ec0563-46f1-45b0-892b-352d088f9517" (UID: "77ec0563-46f1-45b0-892b-352d088f9517"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.398501 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:23:27 crc kubenswrapper[4676]: E0124 00:23:27.399006 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="ceilometer-notification-agent" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.399022 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="ceilometer-notification-agent" Jan 24 00:23:27 crc kubenswrapper[4676]: E0124 00:23:27.399047 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643e6d41-6572-4f21-8651-7f577967bfe8" containerName="glance-httpd" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.399055 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="643e6d41-6572-4f21-8651-7f577967bfe8" containerName="glance-httpd" Jan 24 00:23:27 crc kubenswrapper[4676]: E0124 00:23:27.399071 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77ec0563-46f1-45b0-892b-352d088f9517" containerName="glance-log" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.399079 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="77ec0563-46f1-45b0-892b-352d088f9517" containerName="glance-log" Jan 24 00:23:27 crc kubenswrapper[4676]: E0124 00:23:27.399089 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="proxy-httpd" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.399097 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="proxy-httpd" Jan 24 00:23:27 crc kubenswrapper[4676]: E0124 00:23:27.399116 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="643e6d41-6572-4f21-8651-7f577967bfe8" containerName="glance-log" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.399123 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="643e6d41-6572-4f21-8651-7f577967bfe8" containerName="glance-log" Jan 24 00:23:27 crc kubenswrapper[4676]: E0124 00:23:27.399134 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77ec0563-46f1-45b0-892b-352d088f9517" containerName="glance-httpd" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.399142 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="77ec0563-46f1-45b0-892b-352d088f9517" containerName="glance-httpd" Jan 24 00:23:27 crc kubenswrapper[4676]: E0124 00:23:27.399155 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="sg-core" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.399162 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="sg-core" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.399418 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="sg-core" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.400123 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="ceilometer-notification-agent" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.400155 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="77ec0563-46f1-45b0-892b-352d088f9517" containerName="glance-log" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.400174 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="643e6d41-6572-4f21-8651-7f577967bfe8" containerName="glance-httpd" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.400182 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="643e6d41-6572-4f21-8651-7f577967bfe8" containerName="glance-log" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.400192 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" containerName="proxy-httpd" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.400211 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="77ec0563-46f1-45b0-892b-352d088f9517" containerName="glance-httpd" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.401320 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77ec0563-46f1-45b0-892b-352d088f9517" (UID: "77ec0563-46f1-45b0-892b-352d088f9517"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.402355 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.403797 4676 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.405616 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.410070 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.416668 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.466254 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.466295 4676 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.466316 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.466329 4676 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/77ec0563-46f1-45b0-892b-352d088f9517-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.536498 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-756b7b5794-xhgs6"] Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.537478 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-756b7b5794-xhgs6" podUID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerName="neutron-httpd" containerID="cri-o://ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a" gracePeriod=30 Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.537009 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-756b7b5794-xhgs6" podUID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerName="neutron-api" containerID="cri-o://90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e" gracePeriod=30 Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.569471 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-scripts\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.569532 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6cef518-8385-4315-83d5-f46a6144d5a0-logs\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.569683 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr4ws\" (UniqueName: \"kubernetes.io/projected/a6cef518-8385-4315-83d5-f46a6144d5a0-kube-api-access-sr4ws\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.569743 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.569765 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6cef518-8385-4315-83d5-f46a6144d5a0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.569805 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.569846 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.569870 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-config-data\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.686684 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.686735 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-config-data\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.686775 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-scripts\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.686790 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6cef518-8385-4315-83d5-f46a6144d5a0-logs\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.686873 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr4ws\" (UniqueName: \"kubernetes.io/projected/a6cef518-8385-4315-83d5-f46a6144d5a0-kube-api-access-sr4ws\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.686925 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.686943 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6cef518-8385-4315-83d5-f46a6144d5a0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.686975 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.689959 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.690670 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6cef518-8385-4315-83d5-f46a6144d5a0-logs\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.691082 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6cef518-8385-4315-83d5-f46a6144d5a0-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.691234 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.693597 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-scripts\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.710372 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-config-data\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.716365 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr4ws\" (UniqueName: \"kubernetes.io/projected/a6cef518-8385-4315-83d5-f46a6144d5a0-kube-api-access-sr4ws\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.717572 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6cef518-8385-4315-83d5-f46a6144d5a0-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:27 crc kubenswrapper[4676]: I0124 00:23:27.753926 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"a6cef518-8385-4315-83d5-f46a6144d5a0\") " pod="openstack/glance-default-external-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.025679 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.212829 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd4662bb-5179-4f57-8571-5198dcf69bdb" containerID="0b76bccfcacd7fb1c879fa9b8c0e01b681d270afe951e54be9117c071c170135" exitCode=0 Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.213138 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" event={"ID":"bd4662bb-5179-4f57-8571-5198dcf69bdb","Type":"ContainerDied","Data":"0b76bccfcacd7fb1c879fa9b8c0e01b681d270afe951e54be9117c071c170135"} Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.215163 4676 generic.go:334] "Generic (PLEG): container finished" podID="792cc535-5ed4-4e06-a19c-31ba34c7dfc7" containerID="3ee83666354e8c207624c3af1d827bd9c61811a7455e005083ff4e2e34ef0d81" exitCode=0 Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.215201 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vtxzm" event={"ID":"792cc535-5ed4-4e06-a19c-31ba34c7dfc7","Type":"ContainerDied","Data":"3ee83666354e8c207624c3af1d827bd9c61811a7455e005083ff4e2e34ef0d81"} Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.219486 4676 generic.go:334] "Generic (PLEG): container finished" podID="ee552380-8d96-4a10-a5b8-2deb2e73b15f" containerID="e7fc18fa4adecb5d9d62a155b3da665c4c438818bc9b1324ec05947c45e02af0" exitCode=0 Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.219540 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bpr95" event={"ID":"ee552380-8d96-4a10-a5b8-2deb2e73b15f","Type":"ContainerDied","Data":"e7fc18fa4adecb5d9d62a155b3da665c4c438818bc9b1324ec05947c45e02af0"} Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.232688 4676 generic.go:334] "Generic (PLEG): container finished" podID="7d47faf0-8eaa-474f-8cde-aed6c10f4a05" containerID="cc3b8c014c9450ef0997c6ac2396003671e4bb4fb9c8343ce0debdf9d3796b15" exitCode=0 Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.232784 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a963-account-create-update-jx4g6" event={"ID":"7d47faf0-8eaa-474f-8cde-aed6c10f4a05","Type":"ContainerDied","Data":"cc3b8c014c9450ef0997c6ac2396003671e4bb4fb9c8343ce0debdf9d3796b15"} Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.237327 4676 generic.go:334] "Generic (PLEG): container finished" podID="9cd4cfd3-72ab-45d6-8683-85acb6cadf66" containerID="848f0e8f0de2d0ecec1629bf16e2fb50f2be26fc04335b7c112ae212cbf9035a" exitCode=0 Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.237479 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b7wzs" event={"ID":"9cd4cfd3-72ab-45d6-8683-85acb6cadf66","Type":"ContainerDied","Data":"848f0e8f0de2d0ecec1629bf16e2fb50f2be26fc04335b7c112ae212cbf9035a"} Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.249054 4676 generic.go:334] "Generic (PLEG): container finished" podID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerID="ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a" exitCode=0 Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.249109 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756b7b5794-xhgs6" event={"ID":"1d557c78-075a-44f7-a530-860ae3ec8ffd","Type":"ContainerDied","Data":"ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a"} Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.252723 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"77ec0563-46f1-45b0-892b-352d088f9517","Type":"ContainerDied","Data":"07d4384811e1f2313a45aceab532176f3c3d05348e79e5d4955b534eb2e2868c"} Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.252777 4676 scope.go:117] "RemoveContainer" containerID="1989fb85f0013ac25e4d8fb0c72cdef0a1a08075037f574a25a6701244b8dd50" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.252782 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.275226 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="643e6d41-6572-4f21-8651-7f577967bfe8" path="/var/lib/kubelet/pods/643e6d41-6572-4f21-8651-7f577967bfe8/volumes" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.319122 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.325621 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.345344 4676 scope.go:117] "RemoveContainer" containerID="7c7d07ff8323e9ccead03e4d66a066ad64d46892c577bd47375ca06cbd3b2418" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.360452 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.361912 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.363614 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.363756 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.377066 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.509452 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/042ec835-b8c1-43be-a19d-d70f76128e26-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.509792 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.509812 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.509839 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-config-data\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.509869 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.509888 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h25m4\" (UniqueName: \"kubernetes.io/projected/042ec835-b8c1-43be-a19d-d70f76128e26-kube-api-access-h25m4\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.509926 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/042ec835-b8c1-43be-a19d-d70f76128e26-logs\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.509947 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-scripts\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.612448 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.612492 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h25m4\" (UniqueName: \"kubernetes.io/projected/042ec835-b8c1-43be-a19d-d70f76128e26-kube-api-access-h25m4\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.612539 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/042ec835-b8c1-43be-a19d-d70f76128e26-logs\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.612560 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-scripts\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.612630 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/042ec835-b8c1-43be-a19d-d70f76128e26-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.612669 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.612686 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.612714 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-config-data\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.615745 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/042ec835-b8c1-43be-a19d-d70f76128e26-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.616093 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.617067 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/042ec835-b8c1-43be-a19d-d70f76128e26-logs\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.623588 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-scripts\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.623952 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.626389 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-config-data\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.637362 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/042ec835-b8c1-43be-a19d-d70f76128e26-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.645831 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h25m4\" (UniqueName: \"kubernetes.io/projected/042ec835-b8c1-43be-a19d-d70f76128e26-kube-api-access-h25m4\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.651024 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.671169 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"042ec835-b8c1-43be-a19d-d70f76128e26\") " pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.705699 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.740011 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.920919 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmmfv\" (UniqueName: \"kubernetes.io/projected/2a5f8ec9-ee62-4078-846c-291a47631ffb-kube-api-access-qmmfv\") pod \"2a5f8ec9-ee62-4078-846c-291a47631ffb\" (UID: \"2a5f8ec9-ee62-4078-846c-291a47631ffb\") " Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.921242 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a5f8ec9-ee62-4078-846c-291a47631ffb-operator-scripts\") pod \"2a5f8ec9-ee62-4078-846c-291a47631ffb\" (UID: \"2a5f8ec9-ee62-4078-846c-291a47631ffb\") " Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.925699 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a5f8ec9-ee62-4078-846c-291a47631ffb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2a5f8ec9-ee62-4078-846c-291a47631ffb" (UID: "2a5f8ec9-ee62-4078-846c-291a47631ffb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:28 crc kubenswrapper[4676]: I0124 00:23:28.956318 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a5f8ec9-ee62-4078-846c-291a47631ffb-kube-api-access-qmmfv" (OuterVolumeSpecName: "kube-api-access-qmmfv") pod "2a5f8ec9-ee62-4078-846c-291a47631ffb" (UID: "2a5f8ec9-ee62-4078-846c-291a47631ffb"). InnerVolumeSpecName "kube-api-access-qmmfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.024323 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmmfv\" (UniqueName: \"kubernetes.io/projected/2a5f8ec9-ee62-4078-846c-291a47631ffb-kube-api-access-qmmfv\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.024353 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a5f8ec9-ee62-4078-846c-291a47631ffb-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.263356 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" event={"ID":"2a5f8ec9-ee62-4078-846c-291a47631ffb","Type":"ContainerDied","Data":"e22d8a5b2f69571d42047489f1547c9fe910b6a190e614d344f927ed2bce6dc6"} Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.263424 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e22d8a5b2f69571d42047489f1547c9fe910b6a190e614d344f927ed2bce6dc6" Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.263368 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-95bf-account-create-update-7cvdb" Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.265952 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a6cef518-8385-4315-83d5-f46a6144d5a0","Type":"ContainerStarted","Data":"9b4fae0548ececbcf6b0297fa75fbcf74c89ffcaf56d9df317752603eb9304e3"} Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.356312 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 24 00:23:29 crc kubenswrapper[4676]: W0124 00:23:29.374932 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod042ec835_b8c1_43be_a19d_d70f76128e26.slice/crio-37a7e8266ce7ce32ef66a6acabe1d965b18237492ec2068eb4e79fcb39cf1908 WatchSource:0}: Error finding container 37a7e8266ce7ce32ef66a6acabe1d965b18237492ec2068eb4e79fcb39cf1908: Status 404 returned error can't find the container with id 37a7e8266ce7ce32ef66a6acabe1d965b18237492ec2068eb4e79fcb39cf1908 Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.863126 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.972740 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nljvc\" (UniqueName: \"kubernetes.io/projected/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-kube-api-access-nljvc\") pod \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\" (UID: \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\") " Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.972922 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-operator-scripts\") pod \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\" (UID: \"7d47faf0-8eaa-474f-8cde-aed6c10f4a05\") " Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.973427 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7d47faf0-8eaa-474f-8cde-aed6c10f4a05" (UID: "7d47faf0-8eaa-474f-8cde-aed6c10f4a05"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.975064 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:29 crc kubenswrapper[4676]: I0124 00:23:29.976065 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-kube-api-access-nljvc" (OuterVolumeSpecName: "kube-api-access-nljvc") pod "7d47faf0-8eaa-474f-8cde-aed6c10f4a05" (UID: "7d47faf0-8eaa-474f-8cde-aed6c10f4a05"). InnerVolumeSpecName "kube-api-access-nljvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.077330 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nljvc\" (UniqueName: \"kubernetes.io/projected/7d47faf0-8eaa-474f-8cde-aed6c10f4a05-kube-api-access-nljvc\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.196678 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.282146 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-727rf\" (UniqueName: \"kubernetes.io/projected/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-kube-api-access-727rf\") pod \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\" (UID: \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\") " Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.282256 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-operator-scripts\") pod \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\" (UID: \"792cc535-5ed4-4e06-a19c-31ba34c7dfc7\") " Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.284899 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "792cc535-5ed4-4e06-a19c-31ba34c7dfc7" (UID: "792cc535-5ed4-4e06-a19c-31ba34c7dfc7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.287767 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.310138 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77ec0563-46f1-45b0-892b-352d088f9517" path="/var/lib/kubelet/pods/77ec0563-46f1-45b0-892b-352d088f9517/volumes" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.324762 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-kube-api-access-727rf" (OuterVolumeSpecName: "kube-api-access-727rf") pod "792cc535-5ed4-4e06-a19c-31ba34c7dfc7" (UID: "792cc535-5ed4-4e06-a19c-31ba34c7dfc7"). InnerVolumeSpecName "kube-api-access-727rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.354079 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a6cef518-8385-4315-83d5-f46a6144d5a0","Type":"ContainerStarted","Data":"2a5b6c0202ee01285056feddeb0e637c423f5cffcc6053140ac8681ada0216c3"} Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.354111 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"042ec835-b8c1-43be-a19d-d70f76128e26","Type":"ContainerStarted","Data":"37a7e8266ce7ce32ef66a6acabe1d965b18237492ec2068eb4e79fcb39cf1908"} Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.354125 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" event={"ID":"bd4662bb-5179-4f57-8571-5198dcf69bdb","Type":"ContainerDied","Data":"e0549f129beec80d104c722afe731613b69d0b4983a48a2b5cd879221ad0c238"} Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.354138 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0549f129beec80d104c722afe731613b69d0b4983a48a2b5cd879221ad0c238" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.355731 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.359842 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-vtxzm" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.360073 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-vtxzm" event={"ID":"792cc535-5ed4-4e06-a19c-31ba34c7dfc7","Type":"ContainerDied","Data":"82cc1cf40bad027637ba471334a5a9f3e4778b000f0e88aaab517a858e31e686"} Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.360092 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82cc1cf40bad027637ba471334a5a9f3e4778b000f0e88aaab517a858e31e686" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.375271 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-a963-account-create-update-jx4g6" event={"ID":"7d47faf0-8eaa-474f-8cde-aed6c10f4a05","Type":"ContainerDied","Data":"80f0130f0adbb814673f5191c8d75bf65643388a6bb0829cd96603851a909904"} Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.375311 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80f0130f0adbb814673f5191c8d75bf65643388a6bb0829cd96603851a909904" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.375393 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-a963-account-create-update-jx4g6" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.389730 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-727rf\" (UniqueName: \"kubernetes.io/projected/792cc535-5ed4-4e06-a19c-31ba34c7dfc7-kube-api-access-727rf\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.398950 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.410895 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.491332 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee552380-8d96-4a10-a5b8-2deb2e73b15f-operator-scripts\") pod \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\" (UID: \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\") " Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.493122 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4662bb-5179-4f57-8571-5198dcf69bdb-operator-scripts\") pod \"bd4662bb-5179-4f57-8571-5198dcf69bdb\" (UID: \"bd4662bb-5179-4f57-8571-5198dcf69bdb\") " Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.493174 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wl59\" (UniqueName: \"kubernetes.io/projected/ee552380-8d96-4a10-a5b8-2deb2e73b15f-kube-api-access-2wl59\") pod \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\" (UID: \"ee552380-8d96-4a10-a5b8-2deb2e73b15f\") " Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.493209 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfkb7\" (UniqueName: \"kubernetes.io/projected/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-kube-api-access-cfkb7\") pod \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\" (UID: \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\") " Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.493268 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2ngb\" (UniqueName: \"kubernetes.io/projected/bd4662bb-5179-4f57-8571-5198dcf69bdb-kube-api-access-g2ngb\") pod \"bd4662bb-5179-4f57-8571-5198dcf69bdb\" (UID: \"bd4662bb-5179-4f57-8571-5198dcf69bdb\") " Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.493304 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-operator-scripts\") pod \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\" (UID: \"9cd4cfd3-72ab-45d6-8683-85acb6cadf66\") " Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.493042 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee552380-8d96-4a10-a5b8-2deb2e73b15f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee552380-8d96-4a10-a5b8-2deb2e73b15f" (UID: "ee552380-8d96-4a10-a5b8-2deb2e73b15f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.494669 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee552380-8d96-4a10-a5b8-2deb2e73b15f-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.495162 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9cd4cfd3-72ab-45d6-8683-85acb6cadf66" (UID: "9cd4cfd3-72ab-45d6-8683-85acb6cadf66"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.495644 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd4662bb-5179-4f57-8571-5198dcf69bdb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bd4662bb-5179-4f57-8571-5198dcf69bdb" (UID: "bd4662bb-5179-4f57-8571-5198dcf69bdb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.513492 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-kube-api-access-cfkb7" (OuterVolumeSpecName: "kube-api-access-cfkb7") pod "9cd4cfd3-72ab-45d6-8683-85acb6cadf66" (UID: "9cd4cfd3-72ab-45d6-8683-85acb6cadf66"). InnerVolumeSpecName "kube-api-access-cfkb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.515796 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd4662bb-5179-4f57-8571-5198dcf69bdb-kube-api-access-g2ngb" (OuterVolumeSpecName: "kube-api-access-g2ngb") pod "bd4662bb-5179-4f57-8571-5198dcf69bdb" (UID: "bd4662bb-5179-4f57-8571-5198dcf69bdb"). InnerVolumeSpecName "kube-api-access-g2ngb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.516331 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee552380-8d96-4a10-a5b8-2deb2e73b15f-kube-api-access-2wl59" (OuterVolumeSpecName: "kube-api-access-2wl59") pod "ee552380-8d96-4a10-a5b8-2deb2e73b15f" (UID: "ee552380-8d96-4a10-a5b8-2deb2e73b15f"). InnerVolumeSpecName "kube-api-access-2wl59". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.596187 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bd4662bb-5179-4f57-8571-5198dcf69bdb-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.596206 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wl59\" (UniqueName: \"kubernetes.io/projected/ee552380-8d96-4a10-a5b8-2deb2e73b15f-kube-api-access-2wl59\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.596216 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfkb7\" (UniqueName: \"kubernetes.io/projected/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-kube-api-access-cfkb7\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.596225 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2ngb\" (UniqueName: \"kubernetes.io/projected/bd4662bb-5179-4f57-8571-5198dcf69bdb-kube-api-access-g2ngb\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:30 crc kubenswrapper[4676]: I0124 00:23:30.596235 4676 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9cd4cfd3-72ab-45d6-8683-85acb6cadf66-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.038644 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.215544 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-combined-ca-bundle\") pod \"1d557c78-075a-44f7-a530-860ae3ec8ffd\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.215623 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-httpd-config\") pod \"1d557c78-075a-44f7-a530-860ae3ec8ffd\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.215664 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-ovndb-tls-certs\") pod \"1d557c78-075a-44f7-a530-860ae3ec8ffd\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.215695 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkwcg\" (UniqueName: \"kubernetes.io/projected/1d557c78-075a-44f7-a530-860ae3ec8ffd-kube-api-access-kkwcg\") pod \"1d557c78-075a-44f7-a530-860ae3ec8ffd\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.215749 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-config\") pod \"1d557c78-075a-44f7-a530-860ae3ec8ffd\" (UID: \"1d557c78-075a-44f7-a530-860ae3ec8ffd\") " Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.233454 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "1d557c78-075a-44f7-a530-860ae3ec8ffd" (UID: "1d557c78-075a-44f7-a530-860ae3ec8ffd"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.237396 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d557c78-075a-44f7-a530-860ae3ec8ffd-kube-api-access-kkwcg" (OuterVolumeSpecName: "kube-api-access-kkwcg") pod "1d557c78-075a-44f7-a530-860ae3ec8ffd" (UID: "1d557c78-075a-44f7-a530-860ae3ec8ffd"). InnerVolumeSpecName "kube-api-access-kkwcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.317473 4676 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.317506 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkwcg\" (UniqueName: \"kubernetes.io/projected/1d557c78-075a-44f7-a530-860ae3ec8ffd-kube-api-access-kkwcg\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.330077 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d557c78-075a-44f7-a530-860ae3ec8ffd" (UID: "1d557c78-075a-44f7-a530-860ae3ec8ffd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.368927 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-config" (OuterVolumeSpecName: "config") pod "1d557c78-075a-44f7-a530-860ae3ec8ffd" (UID: "1d557c78-075a-44f7-a530-860ae3ec8ffd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.370531 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "1d557c78-075a-44f7-a530-860ae3ec8ffd" (UID: "1d557c78-075a-44f7-a530-860ae3ec8ffd"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.386972 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-b7wzs" event={"ID":"9cd4cfd3-72ab-45d6-8683-85acb6cadf66","Type":"ContainerDied","Data":"19a2176ef505fcf59da905e70e48985ee4a9f5afc8429b0b5550316e0527bf2e"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.387299 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19a2176ef505fcf59da905e70e48985ee4a9f5afc8429b0b5550316e0527bf2e" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.387398 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-b7wzs" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.399492 4676 generic.go:334] "Generic (PLEG): container finished" podID="ac7dce6b-3bd9-4ad9-9485-83d9384b8bad" containerID="c8c367e4ea3e593fa82e51f67b7f8371ac3844af82364461f69e4513b638a905" exitCode=137 Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.399559 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f876ddf46-fs7qv" event={"ID":"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad","Type":"ContainerDied","Data":"c8c367e4ea3e593fa82e51f67b7f8371ac3844af82364461f69e4513b638a905"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.399585 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f876ddf46-fs7qv" event={"ID":"ac7dce6b-3bd9-4ad9-9485-83d9384b8bad","Type":"ContainerStarted","Data":"7eea686e131fa26ee3ef738923c53a453d90d9eb552a0df5b750dce59c1c117a"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.404993 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a6cef518-8385-4315-83d5-f46a6144d5a0","Type":"ContainerStarted","Data":"d986b5a9792da22ccefff73326e189cc04f8e031e47dca7d8b188a0f58a32bbd"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.406967 4676 generic.go:334] "Generic (PLEG): container finished" podID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerID="90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e" exitCode=0 Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.407022 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756b7b5794-xhgs6" event={"ID":"1d557c78-075a-44f7-a530-860ae3ec8ffd","Type":"ContainerDied","Data":"90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.407047 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-756b7b5794-xhgs6" event={"ID":"1d557c78-075a-44f7-a530-860ae3ec8ffd","Type":"ContainerDied","Data":"2a37de0875eb36fe5f67aef1544f7cf8a8b0cd17d7552f5a5cd3065c861671b8"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.407063 4676 scope.go:117] "RemoveContainer" containerID="ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.407167 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-756b7b5794-xhgs6" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.410001 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"042ec835-b8c1-43be-a19d-d70f76128e26","Type":"ContainerStarted","Data":"b1b68a26c2430d43efd21845e6480121691d5af08b91905ad5de393c697d49d7"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.411221 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bpr95" event={"ID":"ee552380-8d96-4a10-a5b8-2deb2e73b15f","Type":"ContainerDied","Data":"0b915aba37b3b554522a226fb9001f30eb50da19436816f3096df59862d25de5"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.411300 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b915aba37b3b554522a226fb9001f30eb50da19436816f3096df59862d25de5" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.411760 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bpr95" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.416004 4676 generic.go:334] "Generic (PLEG): container finished" podID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerID="9b6a2140f368edfea407714ce9a9afb7bd7057997de5cfab1861b7a16523894e" exitCode=137 Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.416074 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5f9c-account-create-update-9tmf6" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.418308 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf988b4bd-ls7hp" event={"ID":"9d2451a0-4896-46e4-9b9e-e309ccdf02f2","Type":"ContainerDied","Data":"9b6a2140f368edfea407714ce9a9afb7bd7057997de5cfab1861b7a16523894e"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.418360 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf988b4bd-ls7hp" event={"ID":"9d2451a0-4896-46e4-9b9e-e309ccdf02f2","Type":"ContainerStarted","Data":"6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64"} Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.419868 4676 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.419889 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.419898 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d557c78-075a-44f7-a530-860ae3ec8ffd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.449891 4676 scope.go:117] "RemoveContainer" containerID="90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.487179 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.487161361 podStartE2EDuration="4.487161361s" podCreationTimestamp="2026-01-24 00:23:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:31.458550151 +0000 UTC m=+1195.488521152" watchObservedRunningTime="2026-01-24 00:23:31.487161361 +0000 UTC m=+1195.517132352" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.502006 4676 scope.go:117] "RemoveContainer" containerID="ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a" Jan 24 00:23:31 crc kubenswrapper[4676]: E0124 00:23:31.503477 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a\": container with ID starting with ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a not found: ID does not exist" containerID="ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.503504 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a"} err="failed to get container status \"ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a\": rpc error: code = NotFound desc = could not find container \"ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a\": container with ID starting with ac9099c2dcc1c109c19d71ba0aa8ba6b06382c4d9cd67585b0af07f9df929d7a not found: ID does not exist" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.503522 4676 scope.go:117] "RemoveContainer" containerID="90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e" Jan 24 00:23:31 crc kubenswrapper[4676]: E0124 00:23:31.505257 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e\": container with ID starting with 90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e not found: ID does not exist" containerID="90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.505277 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e"} err="failed to get container status \"90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e\": rpc error: code = NotFound desc = could not find container \"90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e\": container with ID starting with 90033ae7fbe939cb7401b9fbf3c1fcdc688cc9fa72836aaba5d889149f24959e not found: ID does not exist" Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.531425 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-756b7b5794-xhgs6"] Jan 24 00:23:31 crc kubenswrapper[4676]: I0124 00:23:31.573862 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-756b7b5794-xhgs6"] Jan 24 00:23:32 crc kubenswrapper[4676]: I0124 00:23:32.264806 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d557c78-075a-44f7-a530-860ae3ec8ffd" path="/var/lib/kubelet/pods/1d557c78-075a-44f7-a530-860ae3ec8ffd/volumes" Jan 24 00:23:32 crc kubenswrapper[4676]: I0124 00:23:32.426485 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"042ec835-b8c1-43be-a19d-d70f76128e26","Type":"ContainerStarted","Data":"40e61576763be5ee0a33a70a8b23ef0b97c5bf95ff7b4040c19eeaaaf7e15abc"} Jan 24 00:23:32 crc kubenswrapper[4676]: I0124 00:23:32.447421 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.4474051150000005 podStartE2EDuration="4.447405115s" podCreationTimestamp="2026-01-24 00:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:23:32.441700931 +0000 UTC m=+1196.471671932" watchObservedRunningTime="2026-01-24 00:23:32.447405115 +0000 UTC m=+1196.477376116" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.437284 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ptgmr"] Jan 24 00:23:35 crc kubenswrapper[4676]: E0124 00:23:35.437900 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd4cfd3-72ab-45d6-8683-85acb6cadf66" containerName="mariadb-database-create" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.437911 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd4cfd3-72ab-45d6-8683-85acb6cadf66" containerName="mariadb-database-create" Jan 24 00:23:35 crc kubenswrapper[4676]: E0124 00:23:35.437923 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerName="neutron-api" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.437932 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerName="neutron-api" Jan 24 00:23:35 crc kubenswrapper[4676]: E0124 00:23:35.437940 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerName="neutron-httpd" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.437946 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerName="neutron-httpd" Jan 24 00:23:35 crc kubenswrapper[4676]: E0124 00:23:35.437960 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792cc535-5ed4-4e06-a19c-31ba34c7dfc7" containerName="mariadb-database-create" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.437965 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="792cc535-5ed4-4e06-a19c-31ba34c7dfc7" containerName="mariadb-database-create" Jan 24 00:23:35 crc kubenswrapper[4676]: E0124 00:23:35.437977 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd4662bb-5179-4f57-8571-5198dcf69bdb" containerName="mariadb-account-create-update" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.437984 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd4662bb-5179-4f57-8571-5198dcf69bdb" containerName="mariadb-account-create-update" Jan 24 00:23:35 crc kubenswrapper[4676]: E0124 00:23:35.437997 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a5f8ec9-ee62-4078-846c-291a47631ffb" containerName="mariadb-account-create-update" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.438003 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a5f8ec9-ee62-4078-846c-291a47631ffb" containerName="mariadb-account-create-update" Jan 24 00:23:35 crc kubenswrapper[4676]: E0124 00:23:35.438017 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d47faf0-8eaa-474f-8cde-aed6c10f4a05" containerName="mariadb-account-create-update" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.438023 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d47faf0-8eaa-474f-8cde-aed6c10f4a05" containerName="mariadb-account-create-update" Jan 24 00:23:35 crc kubenswrapper[4676]: E0124 00:23:35.438034 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee552380-8d96-4a10-a5b8-2deb2e73b15f" containerName="mariadb-database-create" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.438041 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee552380-8d96-4a10-a5b8-2deb2e73b15f" containerName="mariadb-database-create" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.443304 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd4662bb-5179-4f57-8571-5198dcf69bdb" containerName="mariadb-account-create-update" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.443342 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd4cfd3-72ab-45d6-8683-85acb6cadf66" containerName="mariadb-database-create" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.443354 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="792cc535-5ed4-4e06-a19c-31ba34c7dfc7" containerName="mariadb-database-create" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.443367 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerName="neutron-httpd" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.443380 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee552380-8d96-4a10-a5b8-2deb2e73b15f" containerName="mariadb-database-create" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.443402 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a5f8ec9-ee62-4078-846c-291a47631ffb" containerName="mariadb-account-create-update" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.443413 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d557c78-075a-44f7-a530-860ae3ec8ffd" containerName="neutron-api" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.443426 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d47faf0-8eaa-474f-8cde-aed6c10f4a05" containerName="mariadb-account-create-update" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.443983 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.449818 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.450102 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-49ln5" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.450356 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.462674 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ptgmr"] Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.491341 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-scripts\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.491410 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttltl\" (UniqueName: \"kubernetes.io/projected/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-kube-api-access-ttltl\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.491720 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.491783 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-config-data\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.596943 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.596999 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-config-data\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.597056 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-scripts\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.597092 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttltl\" (UniqueName: \"kubernetes.io/projected/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-kube-api-access-ttltl\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.613788 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-scripts\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.616068 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-config-data\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.616187 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.617085 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttltl\" (UniqueName: \"kubernetes.io/projected/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-kube-api-access-ttltl\") pod \"nova-cell0-conductor-db-sync-ptgmr\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:35 crc kubenswrapper[4676]: I0124 00:23:35.773276 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:23:36 crc kubenswrapper[4676]: I0124 00:23:36.299321 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ptgmr"] Jan 24 00:23:36 crc kubenswrapper[4676]: I0124 00:23:36.475755 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ptgmr" event={"ID":"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff","Type":"ContainerStarted","Data":"74f951080123fecb0a6fc99ce221e6ac1980472b308fae5306fc8a030c3089f7"} Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.026030 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.026284 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.063189 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.128157 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.490517 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.490564 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.706812 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.707067 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.748381 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:38 crc kubenswrapper[4676]: I0124 00:23:38.750189 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:39 crc kubenswrapper[4676]: I0124 00:23:39.364156 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:23:39 crc kubenswrapper[4676]: I0124 00:23:39.364206 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:23:39 crc kubenswrapper[4676]: I0124 00:23:39.364246 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:23:39 crc kubenswrapper[4676]: I0124 00:23:39.364953 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f13744ab61f6ff84c30249dfd3e19836649d7bb6c4e4a3db144939c565fd684d"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:23:39 crc kubenswrapper[4676]: I0124 00:23:39.365003 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://f13744ab61f6ff84c30249dfd3e19836649d7bb6c4e4a3db144939c565fd684d" gracePeriod=600 Jan 24 00:23:39 crc kubenswrapper[4676]: I0124 00:23:39.499072 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:39 crc kubenswrapper[4676]: I0124 00:23:39.499368 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.171709 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.172022 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.173534 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.321685 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.321953 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.323850 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f876ddf46-fs7qv" podUID="ac7dce6b-3bd9-4ad9-9485-83d9384b8bad" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.514629 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="f13744ab61f6ff84c30249dfd3e19836649d7bb6c4e4a3db144939c565fd684d" exitCode=0 Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.515439 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"f13744ab61f6ff84c30249dfd3e19836649d7bb6c4e4a3db144939c565fd684d"} Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.515463 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"6c0fdc4fa29c1a85e06a8e0b7d6899a3299af1dbe5c0f08e67bee66057d6c55e"} Jan 24 00:23:40 crc kubenswrapper[4676]: I0124 00:23:40.515479 4676 scope.go:117] "RemoveContainer" containerID="2687110039e3aba350a72ed3647bbafb008d22f301a8b50baa7159c6eca5ba33" Jan 24 00:23:41 crc kubenswrapper[4676]: I0124 00:23:41.388184 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 24 00:23:41 crc kubenswrapper[4676]: I0124 00:23:41.388611 4676 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 24 00:23:41 crc kubenswrapper[4676]: I0124 00:23:41.398204 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 24 00:23:41 crc kubenswrapper[4676]: I0124 00:23:41.577675 4676 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 24 00:23:41 crc kubenswrapper[4676]: I0124 00:23:41.577696 4676 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 24 00:23:42 crc kubenswrapper[4676]: I0124 00:23:42.370126 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:42 crc kubenswrapper[4676]: I0124 00:23:42.587207 4676 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 24 00:23:42 crc kubenswrapper[4676]: I0124 00:23:42.734015 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 24 00:23:49 crc kubenswrapper[4676]: I0124 00:23:49.649560 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ptgmr" event={"ID":"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff","Type":"ContainerStarted","Data":"ba2f94acd9876a58363d3def84609810640e18145307211383bf62958eaadb26"} Jan 24 00:23:49 crc kubenswrapper[4676]: I0124 00:23:49.672429 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ptgmr" podStartSLOduration=2.055133307 podStartE2EDuration="14.672411101s" podCreationTimestamp="2026-01-24 00:23:35 +0000 UTC" firstStartedPulling="2026-01-24 00:23:36.312905834 +0000 UTC m=+1200.342876835" lastFinishedPulling="2026-01-24 00:23:48.930183628 +0000 UTC m=+1212.960154629" observedRunningTime="2026-01-24 00:23:49.664836191 +0000 UTC m=+1213.694807192" watchObservedRunningTime="2026-01-24 00:23:49.672411101 +0000 UTC m=+1213.702382102" Jan 24 00:23:50 crc kubenswrapper[4676]: I0124 00:23:50.173490 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Jan 24 00:23:50 crc kubenswrapper[4676]: I0124 00:23:50.322970 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f876ddf46-fs7qv" podUID="ac7dce6b-3bd9-4ad9-9485-83d9384b8bad" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.149:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.149:8443: connect: connection refused" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.325632 4676 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod54d56910-d4b7-45b1-8699-5af7eaa29b96"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod54d56910-d4b7-45b1-8699-5af7eaa29b96] : Timed out while waiting for systemd to remove kubepods-besteffort-pod54d56910_d4b7_45b1_8699_5af7eaa29b96.slice" Jan 24 00:23:56 crc kubenswrapper[4676]: E0124 00:23:56.326152 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod54d56910-d4b7-45b1-8699-5af7eaa29b96] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod54d56910-d4b7-45b1-8699-5af7eaa29b96] : Timed out while waiting for systemd to remove kubepods-besteffort-pod54d56910_d4b7_45b1_8699_5af7eaa29b96.slice" pod="openstack/ceilometer-0" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.714064 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.799642 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.808512 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.840885 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.843135 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.846234 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.846449 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.858362 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.950834 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.951309 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhdb9\" (UniqueName: \"kubernetes.io/projected/30206279-8386-4575-9457-e76760505e8d-kube-api-access-qhdb9\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.951467 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-log-httpd\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.951602 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-scripts\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.951714 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.951835 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-config-data\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:56 crc kubenswrapper[4676]: I0124 00:23:56.951950 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-run-httpd\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.053266 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-log-httpd\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.053607 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-scripts\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.053678 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-log-httpd\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.053761 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.053844 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-config-data\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.053960 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-run-httpd\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.054254 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-run-httpd\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.054466 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.054597 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhdb9\" (UniqueName: \"kubernetes.io/projected/30206279-8386-4575-9457-e76760505e8d-kube-api-access-qhdb9\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.059900 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-config-data\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.060146 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-scripts\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.060819 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.067224 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.075352 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhdb9\" (UniqueName: \"kubernetes.io/projected/30206279-8386-4575-9457-e76760505e8d-kube-api-access-qhdb9\") pod \"ceilometer-0\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.161532 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.639882 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:23:57 crc kubenswrapper[4676]: W0124 00:23:57.641560 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30206279_8386_4575_9457_e76760505e8d.slice/crio-27bde3b4f936759dfdde3e8fe3259a873617d7354ea2e22a706cd0edd462041d WatchSource:0}: Error finding container 27bde3b4f936759dfdde3e8fe3259a873617d7354ea2e22a706cd0edd462041d: Status 404 returned error can't find the container with id 27bde3b4f936759dfdde3e8fe3259a873617d7354ea2e22a706cd0edd462041d Jan 24 00:23:57 crc kubenswrapper[4676]: I0124 00:23:57.722958 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerStarted","Data":"27bde3b4f936759dfdde3e8fe3259a873617d7354ea2e22a706cd0edd462041d"} Jan 24 00:23:58 crc kubenswrapper[4676]: I0124 00:23:58.266556 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54d56910-d4b7-45b1-8699-5af7eaa29b96" path="/var/lib/kubelet/pods/54d56910-d4b7-45b1-8699-5af7eaa29b96/volumes" Jan 24 00:23:58 crc kubenswrapper[4676]: I0124 00:23:58.732231 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerStarted","Data":"acf37fd15df58f0c480e8ca3019535bc0893909734220600f3f004d89aa3edc9"} Jan 24 00:23:59 crc kubenswrapper[4676]: I0124 00:23:59.741129 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerStarted","Data":"74fc38a94fa794a834f28752c4c0e574b0f6bd303f04b96af63769a91e8933fe"} Jan 24 00:24:00 crc kubenswrapper[4676]: I0124 00:24:00.749646 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerStarted","Data":"993ed8d33422939d102af22915be4b96f85bac159232e43af92e15ee1383a46a"} Jan 24 00:24:01 crc kubenswrapper[4676]: I0124 00:24:01.758240 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerStarted","Data":"cf2f47547ef5fe4c3cd5d6009235ea3a947829e0913a6e11bb70499604aaa440"} Jan 24 00:24:01 crc kubenswrapper[4676]: I0124 00:24:01.758869 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 24 00:24:01 crc kubenswrapper[4676]: I0124 00:24:01.802847 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.008078001 podStartE2EDuration="5.802822488s" podCreationTimestamp="2026-01-24 00:23:56 +0000 UTC" firstStartedPulling="2026-01-24 00:23:57.644403741 +0000 UTC m=+1221.674374742" lastFinishedPulling="2026-01-24 00:24:01.439148228 +0000 UTC m=+1225.469119229" observedRunningTime="2026-01-24 00:24:01.776520659 +0000 UTC m=+1225.806491660" watchObservedRunningTime="2026-01-24 00:24:01.802822488 +0000 UTC m=+1225.832793499" Jan 24 00:24:03 crc kubenswrapper[4676]: I0124 00:24:03.242792 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:24:03 crc kubenswrapper[4676]: I0124 00:24:03.375044 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:24:03 crc kubenswrapper[4676]: I0124 00:24:03.774557 4676 generic.go:334] "Generic (PLEG): container finished" podID="9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff" containerID="ba2f94acd9876a58363d3def84609810640e18145307211383bf62958eaadb26" exitCode=0 Jan 24 00:24:03 crc kubenswrapper[4676]: I0124 00:24:03.774607 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ptgmr" event={"ID":"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff","Type":"ContainerDied","Data":"ba2f94acd9876a58363d3def84609810640e18145307211383bf62958eaadb26"} Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.162033 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.322546 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttltl\" (UniqueName: \"kubernetes.io/projected/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-kube-api-access-ttltl\") pod \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.322585 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-combined-ca-bundle\") pod \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.322668 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-scripts\") pod \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.322707 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-config-data\") pod \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\" (UID: \"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff\") " Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.351615 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-scripts" (OuterVolumeSpecName: "scripts") pod "9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff" (UID: "9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.354505 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-kube-api-access-ttltl" (OuterVolumeSpecName: "kube-api-access-ttltl") pod "9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff" (UID: "9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff"). InnerVolumeSpecName "kube-api-access-ttltl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.398554 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff" (UID: "9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.424305 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttltl\" (UniqueName: \"kubernetes.io/projected/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-kube-api-access-ttltl\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.424328 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.424338 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.426919 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-config-data" (OuterVolumeSpecName: "config-data") pod "9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff" (UID: "9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.525733 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.790957 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ptgmr" event={"ID":"9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff","Type":"ContainerDied","Data":"74f951080123fecb0a6fc99ce221e6ac1980472b308fae5306fc8a030c3089f7"} Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.790993 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74f951080123fecb0a6fc99ce221e6ac1980472b308fae5306fc8a030c3089f7" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.791048 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ptgmr" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.854669 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-f876ddf46-fs7qv" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.874070 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:24:05 crc kubenswrapper[4676]: I0124 00:24:05.983969 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bf988b4bd-ls7hp"] Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.009369 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 24 00:24:06 crc kubenswrapper[4676]: E0124 00:24:06.011076 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff" containerName="nova-cell0-conductor-db-sync" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.011173 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff" containerName="nova-cell0-conductor-db-sync" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.011419 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff" containerName="nova-cell0-conductor-db-sync" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.012085 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.015341 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-49ln5" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.015654 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.028287 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.144213 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/400ba963-913a-401c-8f2e-21005977e0c2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"400ba963-913a-401c-8f2e-21005977e0c2\") " pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.144287 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/400ba963-913a-401c-8f2e-21005977e0c2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"400ba963-913a-401c-8f2e-21005977e0c2\") " pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.144334 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w88s8\" (UniqueName: \"kubernetes.io/projected/400ba963-913a-401c-8f2e-21005977e0c2-kube-api-access-w88s8\") pod \"nova-cell0-conductor-0\" (UID: \"400ba963-913a-401c-8f2e-21005977e0c2\") " pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.246455 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/400ba963-913a-401c-8f2e-21005977e0c2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"400ba963-913a-401c-8f2e-21005977e0c2\") " pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.246513 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/400ba963-913a-401c-8f2e-21005977e0c2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"400ba963-913a-401c-8f2e-21005977e0c2\") " pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.246546 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w88s8\" (UniqueName: \"kubernetes.io/projected/400ba963-913a-401c-8f2e-21005977e0c2-kube-api-access-w88s8\") pod \"nova-cell0-conductor-0\" (UID: \"400ba963-913a-401c-8f2e-21005977e0c2\") " pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.251973 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/400ba963-913a-401c-8f2e-21005977e0c2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"400ba963-913a-401c-8f2e-21005977e0c2\") " pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.265843 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w88s8\" (UniqueName: \"kubernetes.io/projected/400ba963-913a-401c-8f2e-21005977e0c2-kube-api-access-w88s8\") pod \"nova-cell0-conductor-0\" (UID: \"400ba963-913a-401c-8f2e-21005977e0c2\") " pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.266803 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/400ba963-913a-401c-8f2e-21005977e0c2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"400ba963-913a-401c-8f2e-21005977e0c2\") " pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.346515 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.798532 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon-log" containerID="cri-o://ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb" gracePeriod=30 Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.798562 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" containerID="cri-o://6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64" gracePeriod=30 Jan 24 00:24:06 crc kubenswrapper[4676]: I0124 00:24:06.857704 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 24 00:24:07 crc kubenswrapper[4676]: I0124 00:24:07.807541 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"400ba963-913a-401c-8f2e-21005977e0c2","Type":"ContainerStarted","Data":"fb0ae3f2ccbfcd78e945abad4de061cf67626fc297fe90ae5677173d70a3448e"} Jan 24 00:24:07 crc kubenswrapper[4676]: I0124 00:24:07.807880 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"400ba963-913a-401c-8f2e-21005977e0c2","Type":"ContainerStarted","Data":"881e275f0396651176e32e8c7f8fbcde89d91200a6bb57f6d595d6c57c0fa093"} Jan 24 00:24:07 crc kubenswrapper[4676]: I0124 00:24:07.807903 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:07 crc kubenswrapper[4676]: I0124 00:24:07.829668 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.829653971 podStartE2EDuration="2.829653971s" podCreationTimestamp="2026-01-24 00:24:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:07.825766482 +0000 UTC m=+1231.855737483" watchObservedRunningTime="2026-01-24 00:24:07.829653971 +0000 UTC m=+1231.859624972" Jan 24 00:24:10 crc kubenswrapper[4676]: I0124 00:24:10.172480 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Jan 24 00:24:10 crc kubenswrapper[4676]: I0124 00:24:10.831906 4676 generic.go:334] "Generic (PLEG): container finished" podID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerID="6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64" exitCode=0 Jan 24 00:24:10 crc kubenswrapper[4676]: I0124 00:24:10.832010 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf988b4bd-ls7hp" event={"ID":"9d2451a0-4896-46e4-9b9e-e309ccdf02f2","Type":"ContainerDied","Data":"6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64"} Jan 24 00:24:10 crc kubenswrapper[4676]: I0124 00:24:10.832215 4676 scope.go:117] "RemoveContainer" containerID="9b6a2140f368edfea407714ce9a9afb7bd7057997de5cfab1861b7a16523894e" Jan 24 00:24:16 crc kubenswrapper[4676]: I0124 00:24:16.398362 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Jan 24 00:24:16 crc kubenswrapper[4676]: I0124 00:24:16.964817 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-2cwbv"] Jan 24 00:24:16 crc kubenswrapper[4676]: I0124 00:24:16.966867 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:16 crc kubenswrapper[4676]: I0124 00:24:16.973283 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Jan 24 00:24:16 crc kubenswrapper[4676]: I0124 00:24:16.981501 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Jan 24 00:24:16 crc kubenswrapper[4676]: I0124 00:24:16.988638 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2cwbv"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.075585 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-scripts\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.075622 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j96h\" (UniqueName: \"kubernetes.io/projected/8ee94537-f601-4765-95cb-d56518fb7fc6-kube-api-access-2j96h\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.075677 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-config-data\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.075701 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.133395 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.134771 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.141221 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.154283 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.180172 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.180744 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-scripts\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.180762 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j96h\" (UniqueName: \"kubernetes.io/projected/8ee94537-f601-4765-95cb-d56518fb7fc6-kube-api-access-2j96h\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.180813 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-config-data\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.195137 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-config-data\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.196635 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-scripts\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.199098 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.200127 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.202546 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.208920 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.240416 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.253563 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j96h\" (UniqueName: \"kubernetes.io/projected/8ee94537-f601-4765-95cb-d56518fb7fc6-kube-api-access-2j96h\") pod \"nova-cell0-cell-mapping-2cwbv\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.282255 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-config-data\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.282303 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsdvz\" (UniqueName: \"kubernetes.io/projected/0d70bd7f-2410-490a-897f-c17149dbc01d-kube-api-access-fsdvz\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.282341 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.282362 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzqhw\" (UniqueName: \"kubernetes.io/projected/98b86cf1-9245-4a23-aac1-262c54e60716-kube-api-access-pzqhw\") pod \"nova-scheduler-0\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.282473 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d70bd7f-2410-490a-897f-c17149dbc01d-logs\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.282508 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-config-data\") pod \"nova-scheduler-0\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.282550 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.306339 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.388031 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.388093 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-config-data\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.388118 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsdvz\" (UniqueName: \"kubernetes.io/projected/0d70bd7f-2410-490a-897f-c17149dbc01d-kube-api-access-fsdvz\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.388150 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.388172 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzqhw\" (UniqueName: \"kubernetes.io/projected/98b86cf1-9245-4a23-aac1-262c54e60716-kube-api-access-pzqhw\") pod \"nova-scheduler-0\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.388278 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d70bd7f-2410-490a-897f-c17149dbc01d-logs\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.388300 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-config-data\") pod \"nova-scheduler-0\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.395698 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d70bd7f-2410-490a-897f-c17149dbc01d-logs\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.399471 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.412179 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.412286 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-config-data\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.419902 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-config-data\") pod \"nova-scheduler-0\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.498799 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzqhw\" (UniqueName: \"kubernetes.io/projected/98b86cf1-9245-4a23-aac1-262c54e60716-kube-api-access-pzqhw\") pod \"nova-scheduler-0\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.507034 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsdvz\" (UniqueName: \"kubernetes.io/projected/0d70bd7f-2410-490a-897f-c17149dbc01d-kube-api-access-fsdvz\") pod \"nova-api-0\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.546555 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.547810 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.569712 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.580323 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.582751 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.594783 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.618824 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.636336 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.663251 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.700916 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.701157 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-config-data\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.701195 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.701221 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.701239 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9hf5\" (UniqueName: \"kubernetes.io/projected/65324a0f-f62a-4807-85d2-e4b607b2a0b4-kube-api-access-p9hf5\") pod \"nova-cell1-novncproxy-0\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.701259 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f77n\" (UniqueName: \"kubernetes.io/projected/84633021-2bad-4066-a01c-44fef6902524-kube-api-access-8f77n\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.701282 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84633021-2bad-4066-a01c-44fef6902524-logs\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.766607 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.802520 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.802617 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-config-data\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.802666 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.802693 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.802710 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9hf5\" (UniqueName: \"kubernetes.io/projected/65324a0f-f62a-4807-85d2-e4b607b2a0b4-kube-api-access-p9hf5\") pod \"nova-cell1-novncproxy-0\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.802735 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f77n\" (UniqueName: \"kubernetes.io/projected/84633021-2bad-4066-a01c-44fef6902524-kube-api-access-8f77n\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.802756 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84633021-2bad-4066-a01c-44fef6902524-logs\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.803194 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84633021-2bad-4066-a01c-44fef6902524-logs\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.822967 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.828486 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-config-data\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.830220 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.834863 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.846233 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f77n\" (UniqueName: \"kubernetes.io/projected/84633021-2bad-4066-a01c-44fef6902524-kube-api-access-8f77n\") pod \"nova-metadata-0\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.856392 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9hf5\" (UniqueName: \"kubernetes.io/projected/65324a0f-f62a-4807-85d2-e4b607b2a0b4-kube-api-access-p9hf5\") pod \"nova-cell1-novncproxy-0\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.883860 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.904400 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.952418 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-72qp2"] Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.962159 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:17 crc kubenswrapper[4676]: I0124 00:24:17.983020 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-72qp2"] Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.119628 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.119958 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.120000 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.120028 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-svc\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.120051 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2kxc\" (UniqueName: \"kubernetes.io/projected/b8c07196-aa1d-4d14-bc9f-6aec4de13853-kube-api-access-c2kxc\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.120087 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-config\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.223712 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.223780 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.223838 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.223875 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-svc\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.223897 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2kxc\" (UniqueName: \"kubernetes.io/projected/b8c07196-aa1d-4d14-bc9f-6aec4de13853-kube-api-access-c2kxc\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.223946 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-config\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.224839 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-config\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.224839 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.225010 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-svc\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.225350 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.225617 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.254035 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2kxc\" (UniqueName: \"kubernetes.io/projected/b8c07196-aa1d-4d14-bc9f-6aec4de13853-kube-api-access-c2kxc\") pod \"dnsmasq-dns-bccf8f775-72qp2\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.284534 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2cwbv"] Jan 24 00:24:18 crc kubenswrapper[4676]: W0124 00:24:18.308122 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ee94537_f601_4765_95cb_d56518fb7fc6.slice/crio-94c2e3389aa98f25b2feddec05c3db32269606b8ba9977ffa3565e20fb1c9ec4 WatchSource:0}: Error finding container 94c2e3389aa98f25b2feddec05c3db32269606b8ba9977ffa3565e20fb1c9ec4: Status 404 returned error can't find the container with id 94c2e3389aa98f25b2feddec05c3db32269606b8ba9977ffa3565e20fb1c9ec4 Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.352958 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.684081 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.724088 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.729746 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.747320 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.834206 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:24:18 crc kubenswrapper[4676]: W0124 00:24:18.838477 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84633021_2bad_4066_a01c_44fef6902524.slice/crio-6e82aa991fb418472d65f2acfb31fecbeef7f12dbcdb1246d40b5110c0b6ab65 WatchSource:0}: Error finding container 6e82aa991fb418472d65f2acfb31fecbeef7f12dbcdb1246d40b5110c0b6ab65: Status 404 returned error can't find the container with id 6e82aa991fb418472d65f2acfb31fecbeef7f12dbcdb1246d40b5110c0b6ab65 Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.941654 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84633021-2bad-4066-a01c-44fef6902524","Type":"ContainerStarted","Data":"6e82aa991fb418472d65f2acfb31fecbeef7f12dbcdb1246d40b5110c0b6ab65"} Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.955824 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2cwbv" event={"ID":"8ee94537-f601-4765-95cb-d56518fb7fc6","Type":"ContainerStarted","Data":"78da04e8326c020891ea05f676639495efe89b95f4c1821c5af31b8929259b96"} Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.955877 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2cwbv" event={"ID":"8ee94537-f601-4765-95cb-d56518fb7fc6","Type":"ContainerStarted","Data":"94c2e3389aa98f25b2feddec05c3db32269606b8ba9977ffa3565e20fb1c9ec4"} Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.961623 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"65324a0f-f62a-4807-85d2-e4b607b2a0b4","Type":"ContainerStarted","Data":"9d99431ded08302b9fd0007acf0016950fd97f2998e3a74b9630c85127cb16f9"} Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.968561 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"98b86cf1-9245-4a23-aac1-262c54e60716","Type":"ContainerStarted","Data":"842457e0d43257943e9267d87b0c8b3c36220ce75035a13b025aa38f1c36e69d"} Jan 24 00:24:18 crc kubenswrapper[4676]: I0124 00:24:18.990502 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d70bd7f-2410-490a-897f-c17149dbc01d","Type":"ContainerStarted","Data":"db46016d5ba0953d53096b66a815db696cc1c6c9a8457b6a8c4cbfbde714e915"} Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.015459 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fpw6c"] Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.017356 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.022112 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.022497 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.038771 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fpw6c"] Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.067979 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-2cwbv" podStartSLOduration=3.067956781 podStartE2EDuration="3.067956781s" podCreationTimestamp="2026-01-24 00:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:18.995139891 +0000 UTC m=+1243.025110882" watchObservedRunningTime="2026-01-24 00:24:19.067956781 +0000 UTC m=+1243.097927782" Jan 24 00:24:19 crc kubenswrapper[4676]: W0124 00:24:19.076300 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8c07196_aa1d_4d14_bc9f_6aec4de13853.slice/crio-149be033f93bb329597c455a0e6c6a8a209721f648e144a6195c8b2dcfc6602d WatchSource:0}: Error finding container 149be033f93bb329597c455a0e6c6a8a209721f648e144a6195c8b2dcfc6602d: Status 404 returned error can't find the container with id 149be033f93bb329597c455a0e6c6a8a209721f648e144a6195c8b2dcfc6602d Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.081810 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-72qp2"] Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.169467 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2zrw\" (UniqueName: \"kubernetes.io/projected/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-kube-api-access-j2zrw\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.169526 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.169592 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-scripts\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.169667 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-config-data\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.271813 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-config-data\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.272273 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2zrw\" (UniqueName: \"kubernetes.io/projected/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-kube-api-access-j2zrw\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.272360 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.273454 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-scripts\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.277067 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.281366 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-scripts\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.287612 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-config-data\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.297118 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2zrw\" (UniqueName: \"kubernetes.io/projected/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-kube-api-access-j2zrw\") pod \"nova-cell1-conductor-db-sync-fpw6c\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.346509 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:19 crc kubenswrapper[4676]: I0124 00:24:19.916200 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fpw6c"] Jan 24 00:24:19 crc kubenswrapper[4676]: W0124 00:24:19.955564 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1b05e80_51fb_476f_a2b4_bf5a290ea5ae.slice/crio-536f2c9f1d41a95dfb3c6cf31c7557c7581ee9ab704135d4570d8f54edee7545 WatchSource:0}: Error finding container 536f2c9f1d41a95dfb3c6cf31c7557c7581ee9ab704135d4570d8f54edee7545: Status 404 returned error can't find the container with id 536f2c9f1d41a95dfb3c6cf31c7557c7581ee9ab704135d4570d8f54edee7545 Jan 24 00:24:20 crc kubenswrapper[4676]: I0124 00:24:20.010935 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fpw6c" event={"ID":"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae","Type":"ContainerStarted","Data":"536f2c9f1d41a95dfb3c6cf31c7557c7581ee9ab704135d4570d8f54edee7545"} Jan 24 00:24:20 crc kubenswrapper[4676]: I0124 00:24:20.014094 4676 generic.go:334] "Generic (PLEG): container finished" podID="b8c07196-aa1d-4d14-bc9f-6aec4de13853" containerID="ab4cecb7436eb3bdc8075e2778a585be7c7bc22f8b1112d18ec061da1496853f" exitCode=0 Jan 24 00:24:20 crc kubenswrapper[4676]: I0124 00:24:20.014599 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" event={"ID":"b8c07196-aa1d-4d14-bc9f-6aec4de13853","Type":"ContainerDied","Data":"ab4cecb7436eb3bdc8075e2778a585be7c7bc22f8b1112d18ec061da1496853f"} Jan 24 00:24:20 crc kubenswrapper[4676]: I0124 00:24:20.014627 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" event={"ID":"b8c07196-aa1d-4d14-bc9f-6aec4de13853","Type":"ContainerStarted","Data":"149be033f93bb329597c455a0e6c6a8a209721f648e144a6195c8b2dcfc6602d"} Jan 24 00:24:20 crc kubenswrapper[4676]: I0124 00:24:20.174717 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Jan 24 00:24:21 crc kubenswrapper[4676]: I0124 00:24:21.025571 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" event={"ID":"b8c07196-aa1d-4d14-bc9f-6aec4de13853","Type":"ContainerStarted","Data":"2dc83da30d1c8b3f8b6fad2b698ea560b01dbdda3cf6667c49b816b31947008d"} Jan 24 00:24:21 crc kubenswrapper[4676]: I0124 00:24:21.025910 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:21 crc kubenswrapper[4676]: I0124 00:24:21.029098 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fpw6c" event={"ID":"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae","Type":"ContainerStarted","Data":"1e6d788c12343033e8113f7b2adf05ab1c0f4bf8cb67dce88248d068873baa87"} Jan 24 00:24:21 crc kubenswrapper[4676]: I0124 00:24:21.058448 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" podStartSLOduration=4.058420646 podStartE2EDuration="4.058420646s" podCreationTimestamp="2026-01-24 00:24:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:21.044787761 +0000 UTC m=+1245.074758752" watchObservedRunningTime="2026-01-24 00:24:21.058420646 +0000 UTC m=+1245.088391647" Jan 24 00:24:21 crc kubenswrapper[4676]: I0124 00:24:21.066948 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-fpw6c" podStartSLOduration=3.066931495 podStartE2EDuration="3.066931495s" podCreationTimestamp="2026-01-24 00:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:21.064160501 +0000 UTC m=+1245.094131502" watchObservedRunningTime="2026-01-24 00:24:21.066931495 +0000 UTC m=+1245.096902496" Jan 24 00:24:21 crc kubenswrapper[4676]: I0124 00:24:21.637074 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:24:21 crc kubenswrapper[4676]: I0124 00:24:21.664911 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.045079 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84633021-2bad-4066-a01c-44fef6902524","Type":"ContainerStarted","Data":"6301d47d4294644c70fd618b97e56c019270427fbdab822047ac3bfe88e7e594"} Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.045514 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84633021-2bad-4066-a01c-44fef6902524","Type":"ContainerStarted","Data":"b50dc415c434720b0458f8bff74066a1f5105e74beeb732a3d51cdf130cbe5d3"} Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.045637 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="84633021-2bad-4066-a01c-44fef6902524" containerName="nova-metadata-log" containerID="cri-o://b50dc415c434720b0458f8bff74066a1f5105e74beeb732a3d51cdf130cbe5d3" gracePeriod=30 Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.046072 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="84633021-2bad-4066-a01c-44fef6902524" containerName="nova-metadata-metadata" containerID="cri-o://6301d47d4294644c70fd618b97e56c019270427fbdab822047ac3bfe88e7e594" gracePeriod=30 Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.049751 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"65324a0f-f62a-4807-85d2-e4b607b2a0b4","Type":"ContainerStarted","Data":"e4451937240be2f867c73971e845e8f6b7d4b6bee35057ae40b7f593e4b5d73c"} Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.049906 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="65324a0f-f62a-4807-85d2-e4b607b2a0b4" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e4451937240be2f867c73971e845e8f6b7d4b6bee35057ae40b7f593e4b5d73c" gracePeriod=30 Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.062758 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"98b86cf1-9245-4a23-aac1-262c54e60716","Type":"ContainerStarted","Data":"6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997"} Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.064465 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d70bd7f-2410-490a-897f-c17149dbc01d","Type":"ContainerStarted","Data":"d5aab805f55f9b9bf75ccb1416d0d46c71a7f8b8d382ea34d87b6fc4caf8ea0a"} Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.064562 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d70bd7f-2410-490a-897f-c17149dbc01d","Type":"ContainerStarted","Data":"9b214cec1541db1b7c8f4c7ce792f15f1a691f1915cb848b2cef3968d5d483b7"} Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.084643 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.701210031 podStartE2EDuration="6.083112395s" podCreationTimestamp="2026-01-24 00:24:17 +0000 UTC" firstStartedPulling="2026-01-24 00:24:18.850366999 +0000 UTC m=+1242.880338000" lastFinishedPulling="2026-01-24 00:24:22.232269363 +0000 UTC m=+1246.262240364" observedRunningTime="2026-01-24 00:24:23.074702088 +0000 UTC m=+1247.104673089" watchObservedRunningTime="2026-01-24 00:24:23.083112395 +0000 UTC m=+1247.113083396" Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.106180 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.608366912 podStartE2EDuration="6.106163687s" podCreationTimestamp="2026-01-24 00:24:17 +0000 UTC" firstStartedPulling="2026-01-24 00:24:18.736057686 +0000 UTC m=+1242.766028687" lastFinishedPulling="2026-01-24 00:24:22.233854461 +0000 UTC m=+1246.263825462" observedRunningTime="2026-01-24 00:24:23.095531704 +0000 UTC m=+1247.125502705" watchObservedRunningTime="2026-01-24 00:24:23.106163687 +0000 UTC m=+1247.136134688" Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.124557 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.656416536 podStartE2EDuration="6.124538167s" podCreationTimestamp="2026-01-24 00:24:17 +0000 UTC" firstStartedPulling="2026-01-24 00:24:18.773065694 +0000 UTC m=+1242.803036695" lastFinishedPulling="2026-01-24 00:24:22.241187325 +0000 UTC m=+1246.271158326" observedRunningTime="2026-01-24 00:24:23.111507 +0000 UTC m=+1247.141478001" watchObservedRunningTime="2026-01-24 00:24:23.124538167 +0000 UTC m=+1247.154509168" Jan 24 00:24:23 crc kubenswrapper[4676]: I0124 00:24:23.141971 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.632791246 podStartE2EDuration="6.141951858s" podCreationTimestamp="2026-01-24 00:24:17 +0000 UTC" firstStartedPulling="2026-01-24 00:24:18.723898455 +0000 UTC m=+1242.753869456" lastFinishedPulling="2026-01-24 00:24:22.233059067 +0000 UTC m=+1246.263030068" observedRunningTime="2026-01-24 00:24:23.13086436 +0000 UTC m=+1247.160835361" watchObservedRunningTime="2026-01-24 00:24:23.141951858 +0000 UTC m=+1247.171922859" Jan 24 00:24:24 crc kubenswrapper[4676]: I0124 00:24:24.081818 4676 generic.go:334] "Generic (PLEG): container finished" podID="84633021-2bad-4066-a01c-44fef6902524" containerID="b50dc415c434720b0458f8bff74066a1f5105e74beeb732a3d51cdf130cbe5d3" exitCode=143 Jan 24 00:24:24 crc kubenswrapper[4676]: I0124 00:24:24.081961 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84633021-2bad-4066-a01c-44fef6902524","Type":"ContainerDied","Data":"b50dc415c434720b0458f8bff74066a1f5105e74beeb732a3d51cdf130cbe5d3"} Jan 24 00:24:27 crc kubenswrapper[4676]: I0124 00:24:27.166573 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 24 00:24:27 crc kubenswrapper[4676]: I0124 00:24:27.620218 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 24 00:24:27 crc kubenswrapper[4676]: I0124 00:24:27.620267 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 24 00:24:27 crc kubenswrapper[4676]: I0124 00:24:27.651353 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 24 00:24:27 crc kubenswrapper[4676]: I0124 00:24:27.767561 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 24 00:24:27 crc kubenswrapper[4676]: I0124 00:24:27.767608 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 24 00:24:27 crc kubenswrapper[4676]: I0124 00:24:27.884727 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:27 crc kubenswrapper[4676]: I0124 00:24:27.905014 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 24 00:24:27 crc kubenswrapper[4676]: I0124 00:24:27.905055 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.118360 4676 generic.go:334] "Generic (PLEG): container finished" podID="8ee94537-f601-4765-95cb-d56518fb7fc6" containerID="78da04e8326c020891ea05f676639495efe89b95f4c1821c5af31b8929259b96" exitCode=0 Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.118454 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2cwbv" event={"ID":"8ee94537-f601-4765-95cb-d56518fb7fc6","Type":"ContainerDied","Data":"78da04e8326c020891ea05f676639495efe89b95f4c1821c5af31b8929259b96"} Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.120526 4676 generic.go:334] "Generic (PLEG): container finished" podID="a1b05e80-51fb-476f-a2b4-bf5a290ea5ae" containerID="1e6d788c12343033e8113f7b2adf05ab1c0f4bf8cb67dce88248d068873baa87" exitCode=0 Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.121415 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fpw6c" event={"ID":"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae","Type":"ContainerDied","Data":"1e6d788c12343033e8113f7b2adf05ab1c0f4bf8cb67dce88248d068873baa87"} Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.165666 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.356305 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.410465 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-jbptt"] Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.410960 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" podUID="9b4acc7d-6a5f-4544-b73e-61b9a620c17d" containerName="dnsmasq-dns" containerID="cri-o://3c1cdcc2ffd783555c7e0583ecb6ac91d2d0c15034969f7f5f6f8dba79f95867" gracePeriod=10 Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.884610 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 24 00:24:28 crc kubenswrapper[4676]: I0124 00:24:28.884855 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.129778 4676 generic.go:334] "Generic (PLEG): container finished" podID="9b4acc7d-6a5f-4544-b73e-61b9a620c17d" containerID="3c1cdcc2ffd783555c7e0583ecb6ac91d2d0c15034969f7f5f6f8dba79f95867" exitCode=0 Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.130297 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" event={"ID":"9b4acc7d-6a5f-4544-b73e-61b9a620c17d","Type":"ContainerDied","Data":"3c1cdcc2ffd783555c7e0583ecb6ac91d2d0c15034969f7f5f6f8dba79f95867"} Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.130629 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" event={"ID":"9b4acc7d-6a5f-4544-b73e-61b9a620c17d","Type":"ContainerDied","Data":"eec257beeff7ba3f585f1d2f0b787abd4407c78ad654a5cbdd01eb786faa02e2"} Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.130777 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eec257beeff7ba3f585f1d2f0b787abd4407c78ad654a5cbdd01eb786faa02e2" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.189771 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.273057 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-config\") pod \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.273148 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-swift-storage-0\") pod \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.273302 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-sb\") pod \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.273342 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-svc\") pod \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.273465 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxx64\" (UniqueName: \"kubernetes.io/projected/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-kube-api-access-xxx64\") pod \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.273486 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-nb\") pod \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\" (UID: \"9b4acc7d-6a5f-4544-b73e-61b9a620c17d\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.308503 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-kube-api-access-xxx64" (OuterVolumeSpecName: "kube-api-access-xxx64") pod "9b4acc7d-6a5f-4544-b73e-61b9a620c17d" (UID: "9b4acc7d-6a5f-4544-b73e-61b9a620c17d"). InnerVolumeSpecName "kube-api-access-xxx64". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.375647 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxx64\" (UniqueName: \"kubernetes.io/projected/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-kube-api-access-xxx64\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.393480 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9b4acc7d-6a5f-4544-b73e-61b9a620c17d" (UID: "9b4acc7d-6a5f-4544-b73e-61b9a620c17d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.394849 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9b4acc7d-6a5f-4544-b73e-61b9a620c17d" (UID: "9b4acc7d-6a5f-4544-b73e-61b9a620c17d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.401217 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-config" (OuterVolumeSpecName: "config") pod "9b4acc7d-6a5f-4544-b73e-61b9a620c17d" (UID: "9b4acc7d-6a5f-4544-b73e-61b9a620c17d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.411795 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9b4acc7d-6a5f-4544-b73e-61b9a620c17d" (UID: "9b4acc7d-6a5f-4544-b73e-61b9a620c17d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.412758 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9b4acc7d-6a5f-4544-b73e-61b9a620c17d" (UID: "9b4acc7d-6a5f-4544-b73e-61b9a620c17d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.481221 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.481243 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.481252 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.481261 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.481269 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9b4acc7d-6a5f-4544-b73e-61b9a620c17d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.731692 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.736866 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.786837 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-scripts\") pod \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.786887 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-combined-ca-bundle\") pod \"8ee94537-f601-4765-95cb-d56518fb7fc6\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.786967 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-config-data\") pod \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.787050 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2zrw\" (UniqueName: \"kubernetes.io/projected/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-kube-api-access-j2zrw\") pod \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.787107 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-scripts\") pod \"8ee94537-f601-4765-95cb-d56518fb7fc6\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.787124 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-combined-ca-bundle\") pod \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\" (UID: \"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.787167 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-config-data\") pod \"8ee94537-f601-4765-95cb-d56518fb7fc6\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.787242 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j96h\" (UniqueName: \"kubernetes.io/projected/8ee94537-f601-4765-95cb-d56518fb7fc6-kube-api-access-2j96h\") pod \"8ee94537-f601-4765-95cb-d56518fb7fc6\" (UID: \"8ee94537-f601-4765-95cb-d56518fb7fc6\") " Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.798273 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ee94537-f601-4765-95cb-d56518fb7fc6-kube-api-access-2j96h" (OuterVolumeSpecName: "kube-api-access-2j96h") pod "8ee94537-f601-4765-95cb-d56518fb7fc6" (UID: "8ee94537-f601-4765-95cb-d56518fb7fc6"). InnerVolumeSpecName "kube-api-access-2j96h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.809674 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-scripts" (OuterVolumeSpecName: "scripts") pod "8ee94537-f601-4765-95cb-d56518fb7fc6" (UID: "8ee94537-f601-4765-95cb-d56518fb7fc6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.814159 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-kube-api-access-j2zrw" (OuterVolumeSpecName: "kube-api-access-j2zrw") pod "a1b05e80-51fb-476f-a2b4-bf5a290ea5ae" (UID: "a1b05e80-51fb-476f-a2b4-bf5a290ea5ae"). InnerVolumeSpecName "kube-api-access-j2zrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.821713 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-scripts" (OuterVolumeSpecName: "scripts") pod "a1b05e80-51fb-476f-a2b4-bf5a290ea5ae" (UID: "a1b05e80-51fb-476f-a2b4-bf5a290ea5ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.833456 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1b05e80-51fb-476f-a2b4-bf5a290ea5ae" (UID: "a1b05e80-51fb-476f-a2b4-bf5a290ea5ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.836506 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-config-data" (OuterVolumeSpecName: "config-data") pod "a1b05e80-51fb-476f-a2b4-bf5a290ea5ae" (UID: "a1b05e80-51fb-476f-a2b4-bf5a290ea5ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.839603 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ee94537-f601-4765-95cb-d56518fb7fc6" (UID: "8ee94537-f601-4765-95cb-d56518fb7fc6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.854276 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-config-data" (OuterVolumeSpecName: "config-data") pod "8ee94537-f601-4765-95cb-d56518fb7fc6" (UID: "8ee94537-f601-4765-95cb-d56518fb7fc6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.889319 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2zrw\" (UniqueName: \"kubernetes.io/projected/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-kube-api-access-j2zrw\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.889349 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.889358 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.889367 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.889389 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j96h\" (UniqueName: \"kubernetes.io/projected/8ee94537-f601-4765-95cb-d56518fb7fc6-kube-api-access-2j96h\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.889398 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.889407 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94537-f601-4765-95cb-d56518fb7fc6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:29 crc kubenswrapper[4676]: I0124 00:24:29.889416 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.139423 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2cwbv" event={"ID":"8ee94537-f601-4765-95cb-d56518fb7fc6","Type":"ContainerDied","Data":"94c2e3389aa98f25b2feddec05c3db32269606b8ba9977ffa3565e20fb1c9ec4"} Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.139475 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94c2e3389aa98f25b2feddec05c3db32269606b8ba9977ffa3565e20fb1c9ec4" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.139490 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2cwbv" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.141343 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-jbptt" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.141352 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-fpw6c" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.141402 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-fpw6c" event={"ID":"a1b05e80-51fb-476f-a2b4-bf5a290ea5ae","Type":"ContainerDied","Data":"536f2c9f1d41a95dfb3c6cf31c7557c7581ee9ab704135d4570d8f54edee7545"} Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.141442 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="536f2c9f1d41a95dfb3c6cf31c7557c7581ee9ab704135d4570d8f54edee7545" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.181104 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-jbptt"] Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.181261 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-bf988b4bd-ls7hp" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.181342 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.199682 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-jbptt"] Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.266334 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b4acc7d-6a5f-4544-b73e-61b9a620c17d" path="/var/lib/kubelet/pods/9b4acc7d-6a5f-4544-b73e-61b9a620c17d/volumes" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.293803 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 24 00:24:30 crc kubenswrapper[4676]: E0124 00:24:30.294247 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4acc7d-6a5f-4544-b73e-61b9a620c17d" containerName="dnsmasq-dns" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.294262 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4acc7d-6a5f-4544-b73e-61b9a620c17d" containerName="dnsmasq-dns" Jan 24 00:24:30 crc kubenswrapper[4676]: E0124 00:24:30.294275 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee94537-f601-4765-95cb-d56518fb7fc6" containerName="nova-manage" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.294284 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee94537-f601-4765-95cb-d56518fb7fc6" containerName="nova-manage" Jan 24 00:24:30 crc kubenswrapper[4676]: E0124 00:24:30.294303 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b05e80-51fb-476f-a2b4-bf5a290ea5ae" containerName="nova-cell1-conductor-db-sync" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.294309 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b05e80-51fb-476f-a2b4-bf5a290ea5ae" containerName="nova-cell1-conductor-db-sync" Jan 24 00:24:30 crc kubenswrapper[4676]: E0124 00:24:30.294335 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b4acc7d-6a5f-4544-b73e-61b9a620c17d" containerName="init" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.294341 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b4acc7d-6a5f-4544-b73e-61b9a620c17d" containerName="init" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.294515 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b4acc7d-6a5f-4544-b73e-61b9a620c17d" containerName="dnsmasq-dns" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.294528 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b05e80-51fb-476f-a2b4-bf5a290ea5ae" containerName="nova-cell1-conductor-db-sync" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.294543 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ee94537-f601-4765-95cb-d56518fb7fc6" containerName="nova-manage" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.295143 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.297674 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.303066 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.399416 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4420d197-3908-406b-9661-67d64ccd7768-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4420d197-3908-406b-9661-67d64ccd7768\") " pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.399498 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4420d197-3908-406b-9661-67d64ccd7768-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4420d197-3908-406b-9661-67d64ccd7768\") " pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.399535 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgxh6\" (UniqueName: \"kubernetes.io/projected/4420d197-3908-406b-9661-67d64ccd7768-kube-api-access-cgxh6\") pod \"nova-cell1-conductor-0\" (UID: \"4420d197-3908-406b-9661-67d64ccd7768\") " pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.424942 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.425154 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-log" containerID="cri-o://9b214cec1541db1b7c8f4c7ce792f15f1a691f1915cb848b2cef3968d5d483b7" gracePeriod=30 Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.425460 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-api" containerID="cri-o://d5aab805f55f9b9bf75ccb1416d0d46c71a7f8b8d382ea34d87b6fc4caf8ea0a" gracePeriod=30 Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.447979 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.448172 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="98b86cf1-9245-4a23-aac1-262c54e60716" containerName="nova-scheduler-scheduler" containerID="cri-o://6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997" gracePeriod=30 Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.500763 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgxh6\" (UniqueName: \"kubernetes.io/projected/4420d197-3908-406b-9661-67d64ccd7768-kube-api-access-cgxh6\") pod \"nova-cell1-conductor-0\" (UID: \"4420d197-3908-406b-9661-67d64ccd7768\") " pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.501203 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4420d197-3908-406b-9661-67d64ccd7768-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4420d197-3908-406b-9661-67d64ccd7768\") " pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.501269 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4420d197-3908-406b-9661-67d64ccd7768-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4420d197-3908-406b-9661-67d64ccd7768\") " pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.505344 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4420d197-3908-406b-9661-67d64ccd7768-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4420d197-3908-406b-9661-67d64ccd7768\") " pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.506326 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4420d197-3908-406b-9661-67d64ccd7768-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4420d197-3908-406b-9661-67d64ccd7768\") " pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.527291 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgxh6\" (UniqueName: \"kubernetes.io/projected/4420d197-3908-406b-9661-67d64ccd7768-kube-api-access-cgxh6\") pod \"nova-cell1-conductor-0\" (UID: \"4420d197-3908-406b-9661-67d64ccd7768\") " pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:30 crc kubenswrapper[4676]: I0124 00:24:30.691680 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:31 crc kubenswrapper[4676]: I0124 00:24:31.150431 4676 generic.go:334] "Generic (PLEG): container finished" podID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerID="9b214cec1541db1b7c8f4c7ce792f15f1a691f1915cb848b2cef3968d5d483b7" exitCode=143 Jan 24 00:24:31 crc kubenswrapper[4676]: I0124 00:24:31.150507 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d70bd7f-2410-490a-897f-c17149dbc01d","Type":"ContainerDied","Data":"9b214cec1541db1b7c8f4c7ce792f15f1a691f1915cb848b2cef3968d5d483b7"} Jan 24 00:24:31 crc kubenswrapper[4676]: I0124 00:24:31.233137 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 24 00:24:32 crc kubenswrapper[4676]: I0124 00:24:32.158604 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4420d197-3908-406b-9661-67d64ccd7768","Type":"ContainerStarted","Data":"7adaaa21bfe2214008bad41a454b7f4029447556182d11dec2c3d953ee0b1a9d"} Jan 24 00:24:32 crc kubenswrapper[4676]: I0124 00:24:32.158854 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4420d197-3908-406b-9661-67d64ccd7768","Type":"ContainerStarted","Data":"622536c9ee0ca1723858277205b67fe8775384dce8b6bb05d90cf3dd2d1ed2d6"} Jan 24 00:24:32 crc kubenswrapper[4676]: I0124 00:24:32.159039 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:32 crc kubenswrapper[4676]: E0124 00:24:32.632463 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 24 00:24:32 crc kubenswrapper[4676]: E0124 00:24:32.638222 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 24 00:24:32 crc kubenswrapper[4676]: E0124 00:24:32.650338 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 24 00:24:32 crc kubenswrapper[4676]: E0124 00:24:32.650463 4676 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="98b86cf1-9245-4a23-aac1-262c54e60716" containerName="nova-scheduler-scheduler" Jan 24 00:24:33 crc kubenswrapper[4676]: I0124 00:24:33.067612 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.067593391 podStartE2EDuration="3.067593391s" podCreationTimestamp="2026-01-24 00:24:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:32.171664195 +0000 UTC m=+1256.201635196" watchObservedRunningTime="2026-01-24 00:24:33.067593391 +0000 UTC m=+1257.097564382" Jan 24 00:24:33 crc kubenswrapper[4676]: I0124 00:24:33.074231 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 24 00:24:33 crc kubenswrapper[4676]: I0124 00:24:33.074439 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c626d726-d406-41c7-8e66-07cc148d3aa7" containerName="kube-state-metrics" containerID="cri-o://0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62" gracePeriod=30 Jan 24 00:24:33 crc kubenswrapper[4676]: I0124 00:24:33.595149 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 24 00:24:33 crc kubenswrapper[4676]: I0124 00:24:33.656657 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vqg2d\" (UniqueName: \"kubernetes.io/projected/c626d726-d406-41c7-8e66-07cc148d3aa7-kube-api-access-vqg2d\") pod \"c626d726-d406-41c7-8e66-07cc148d3aa7\" (UID: \"c626d726-d406-41c7-8e66-07cc148d3aa7\") " Jan 24 00:24:33 crc kubenswrapper[4676]: I0124 00:24:33.702005 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c626d726-d406-41c7-8e66-07cc148d3aa7-kube-api-access-vqg2d" (OuterVolumeSpecName: "kube-api-access-vqg2d") pod "c626d726-d406-41c7-8e66-07cc148d3aa7" (UID: "c626d726-d406-41c7-8e66-07cc148d3aa7"). InnerVolumeSpecName "kube-api-access-vqg2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:33 crc kubenswrapper[4676]: I0124 00:24:33.775695 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vqg2d\" (UniqueName: \"kubernetes.io/projected/c626d726-d406-41c7-8e66-07cc148d3aa7-kube-api-access-vqg2d\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.175410 4676 generic.go:334] "Generic (PLEG): container finished" podID="c626d726-d406-41c7-8e66-07cc148d3aa7" containerID="0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62" exitCode=2 Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.175696 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c626d726-d406-41c7-8e66-07cc148d3aa7","Type":"ContainerDied","Data":"0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62"} Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.175721 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c626d726-d406-41c7-8e66-07cc148d3aa7","Type":"ContainerDied","Data":"a0d93a838adacca6f5b952cbfd2b7bfc9bf0b2ee8a2c8c9cf055df1b2e54dbca"} Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.175736 4676 scope.go:117] "RemoveContainer" containerID="0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.175863 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.202480 4676 scope.go:117] "RemoveContainer" containerID="0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62" Jan 24 00:24:34 crc kubenswrapper[4676]: E0124 00:24:34.203336 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62\": container with ID starting with 0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62 not found: ID does not exist" containerID="0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.203392 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62"} err="failed to get container status \"0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62\": rpc error: code = NotFound desc = could not find container \"0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62\": container with ID starting with 0c6072b7b64b06470b62859afd989687c08854e4403eafaa80685de7b9b6ba62 not found: ID does not exist" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.207836 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.216720 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.229416 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 24 00:24:34 crc kubenswrapper[4676]: E0124 00:24:34.229994 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c626d726-d406-41c7-8e66-07cc148d3aa7" containerName="kube-state-metrics" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.230061 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="c626d726-d406-41c7-8e66-07cc148d3aa7" containerName="kube-state-metrics" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.230300 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="c626d726-d406-41c7-8e66-07cc148d3aa7" containerName="kube-state-metrics" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.231323 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.234081 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.234405 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.239883 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.274959 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c626d726-d406-41c7-8e66-07cc148d3aa7" path="/var/lib/kubelet/pods/c626d726-d406-41c7-8e66-07cc148d3aa7/volumes" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.285455 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.285497 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.285516 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.285570 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjsqf\" (UniqueName: \"kubernetes.io/projected/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-kube-api-access-pjsqf\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.387993 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.388220 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.388303 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.388404 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjsqf\" (UniqueName: \"kubernetes.io/projected/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-kube-api-access-pjsqf\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.393440 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.393810 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.406328 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.410176 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjsqf\" (UniqueName: \"kubernetes.io/projected/25ee749c-0b84-4abd-9fe0-a6f23c0c912d-kube-api-access-pjsqf\") pod \"kube-state-metrics-0\" (UID: \"25ee749c-0b84-4abd-9fe0-a6f23c0c912d\") " pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.601499 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.734972 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.795631 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzqhw\" (UniqueName: \"kubernetes.io/projected/98b86cf1-9245-4a23-aac1-262c54e60716-kube-api-access-pzqhw\") pod \"98b86cf1-9245-4a23-aac1-262c54e60716\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.795955 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-config-data\") pod \"98b86cf1-9245-4a23-aac1-262c54e60716\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.796147 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-combined-ca-bundle\") pod \"98b86cf1-9245-4a23-aac1-262c54e60716\" (UID: \"98b86cf1-9245-4a23-aac1-262c54e60716\") " Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.814904 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98b86cf1-9245-4a23-aac1-262c54e60716-kube-api-access-pzqhw" (OuterVolumeSpecName: "kube-api-access-pzqhw") pod "98b86cf1-9245-4a23-aac1-262c54e60716" (UID: "98b86cf1-9245-4a23-aac1-262c54e60716"). InnerVolumeSpecName "kube-api-access-pzqhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.875672 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-config-data" (OuterVolumeSpecName: "config-data") pod "98b86cf1-9245-4a23-aac1-262c54e60716" (UID: "98b86cf1-9245-4a23-aac1-262c54e60716"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.877551 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98b86cf1-9245-4a23-aac1-262c54e60716" (UID: "98b86cf1-9245-4a23-aac1-262c54e60716"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.899794 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzqhw\" (UniqueName: \"kubernetes.io/projected/98b86cf1-9245-4a23-aac1-262c54e60716-kube-api-access-pzqhw\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.899823 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:34 crc kubenswrapper[4676]: I0124 00:24:34.899833 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b86cf1-9245-4a23-aac1-262c54e60716-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.183336 4676 generic.go:334] "Generic (PLEG): container finished" podID="98b86cf1-9245-4a23-aac1-262c54e60716" containerID="6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997" exitCode=0 Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.183558 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"98b86cf1-9245-4a23-aac1-262c54e60716","Type":"ContainerDied","Data":"6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997"} Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.183643 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"98b86cf1-9245-4a23-aac1-262c54e60716","Type":"ContainerDied","Data":"842457e0d43257943e9267d87b0c8b3c36220ce75035a13b025aa38f1c36e69d"} Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.183650 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.183693 4676 scope.go:117] "RemoveContainer" containerID="6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.202805 4676 scope.go:117] "RemoveContainer" containerID="6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997" Jan 24 00:24:35 crc kubenswrapper[4676]: E0124 00:24:35.203598 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997\": container with ID starting with 6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997 not found: ID does not exist" containerID="6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.203622 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997"} err="failed to get container status \"6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997\": rpc error: code = NotFound desc = could not find container \"6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997\": container with ID starting with 6fd839294bba279baff5cde259e88fab5755b95b8b906e8844a646f2f9e0e997 not found: ID does not exist" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.217152 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.224158 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.242168 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:24:35 crc kubenswrapper[4676]: E0124 00:24:35.242526 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b86cf1-9245-4a23-aac1-262c54e60716" containerName="nova-scheduler-scheduler" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.242542 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b86cf1-9245-4a23-aac1-262c54e60716" containerName="nova-scheduler-scheduler" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.242718 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="98b86cf1-9245-4a23-aac1-262c54e60716" containerName="nova-scheduler-scheduler" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.243256 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.250055 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.267823 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.316405 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.316727 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vr9g\" (UniqueName: \"kubernetes.io/projected/32d678d8-bc5d-4993-afc3-c6700d53b00e-kube-api-access-5vr9g\") pod \"nova-scheduler-0\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.316880 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-config-data\") pod \"nova-scheduler-0\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.324692 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.421513 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.421935 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vr9g\" (UniqueName: \"kubernetes.io/projected/32d678d8-bc5d-4993-afc3-c6700d53b00e-kube-api-access-5vr9g\") pod \"nova-scheduler-0\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.421968 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-config-data\") pod \"nova-scheduler-0\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.429999 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.430270 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-config-data\") pod \"nova-scheduler-0\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.445011 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vr9g\" (UniqueName: \"kubernetes.io/projected/32d678d8-bc5d-4993-afc3-c6700d53b00e-kube-api-access-5vr9g\") pod \"nova-scheduler-0\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " pod="openstack/nova-scheduler-0" Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.547701 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.547948 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="ceilometer-central-agent" containerID="cri-o://acf37fd15df58f0c480e8ca3019535bc0893909734220600f3f004d89aa3edc9" gracePeriod=30 Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.548011 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="proxy-httpd" containerID="cri-o://cf2f47547ef5fe4c3cd5d6009235ea3a947829e0913a6e11bb70499604aaa440" gracePeriod=30 Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.548031 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="sg-core" containerID="cri-o://993ed8d33422939d102af22915be4b96f85bac159232e43af92e15ee1383a46a" gracePeriod=30 Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.548104 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="ceilometer-notification-agent" containerID="cri-o://74fc38a94fa794a834f28752c4c0e574b0f6bd303f04b96af63769a91e8933fe" gracePeriod=30 Jan 24 00:24:35 crc kubenswrapper[4676]: I0124 00:24:35.646517 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.169256 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.211191 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.213777 4676 generic.go:334] "Generic (PLEG): container finished" podID="30206279-8386-4575-9457-e76760505e8d" containerID="cf2f47547ef5fe4c3cd5d6009235ea3a947829e0913a6e11bb70499604aaa440" exitCode=0 Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.213827 4676 generic.go:334] "Generic (PLEG): container finished" podID="30206279-8386-4575-9457-e76760505e8d" containerID="993ed8d33422939d102af22915be4b96f85bac159232e43af92e15ee1383a46a" exitCode=2 Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.213837 4676 generic.go:334] "Generic (PLEG): container finished" podID="30206279-8386-4575-9457-e76760505e8d" containerID="acf37fd15df58f0c480e8ca3019535bc0893909734220600f3f004d89aa3edc9" exitCode=0 Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.213878 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerDied","Data":"cf2f47547ef5fe4c3cd5d6009235ea3a947829e0913a6e11bb70499604aaa440"} Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.213930 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerDied","Data":"993ed8d33422939d102af22915be4b96f85bac159232e43af92e15ee1383a46a"} Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.213939 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerDied","Data":"acf37fd15df58f0c480e8ca3019535bc0893909734220600f3f004d89aa3edc9"} Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.224533 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"32d678d8-bc5d-4993-afc3-c6700d53b00e","Type":"ContainerStarted","Data":"d01bc4c049fe259a880c270ae5608bf35ba2c89a0a3367424a3555670349c800"} Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.249707 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25ee749c-0b84-4abd-9fe0-a6f23c0c912d","Type":"ContainerStarted","Data":"46aba89aa56be268b7b5a98dfb66b1b75ba3d5f9d305baac12a530ef54c5a879"} Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.249752 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"25ee749c-0b84-4abd-9fe0-a6f23c0c912d","Type":"ContainerStarted","Data":"4957534fb4d33c9d7ee550c51f7102ddf197f85e71897067791549e65d161c2a"} Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.251113 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.296719 4676 generic.go:334] "Generic (PLEG): container finished" podID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerID="d5aab805f55f9b9bf75ccb1416d0d46c71a7f8b8d382ea34d87b6fc4caf8ea0a" exitCode=0 Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.296960 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.340397 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98b86cf1-9245-4a23-aac1-262c54e60716" path="/var/lib/kubelet/pods/98b86cf1-9245-4a23-aac1-262c54e60716/volumes" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.343502 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-combined-ca-bundle\") pod \"0d70bd7f-2410-490a-897f-c17149dbc01d\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.343579 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d70bd7f-2410-490a-897f-c17149dbc01d-logs\") pod \"0d70bd7f-2410-490a-897f-c17149dbc01d\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.343727 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-config-data\") pod \"0d70bd7f-2410-490a-897f-c17149dbc01d\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.343967 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsdvz\" (UniqueName: \"kubernetes.io/projected/0d70bd7f-2410-490a-897f-c17149dbc01d-kube-api-access-fsdvz\") pod \"0d70bd7f-2410-490a-897f-c17149dbc01d\" (UID: \"0d70bd7f-2410-490a-897f-c17149dbc01d\") " Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.344392 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d70bd7f-2410-490a-897f-c17149dbc01d-logs" (OuterVolumeSpecName: "logs") pod "0d70bd7f-2410-490a-897f-c17149dbc01d" (UID: "0d70bd7f-2410-490a-897f-c17149dbc01d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.344945 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d70bd7f-2410-490a-897f-c17149dbc01d-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.377364 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d70bd7f-2410-490a-897f-c17149dbc01d-kube-api-access-fsdvz" (OuterVolumeSpecName: "kube-api-access-fsdvz") pod "0d70bd7f-2410-490a-897f-c17149dbc01d" (UID: "0d70bd7f-2410-490a-897f-c17149dbc01d"). InnerVolumeSpecName "kube-api-access-fsdvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.385755 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-config-data" (OuterVolumeSpecName: "config-data") pod "0d70bd7f-2410-490a-897f-c17149dbc01d" (UID: "0d70bd7f-2410-490a-897f-c17149dbc01d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.397012 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.982400977 podStartE2EDuration="2.396991283s" podCreationTimestamp="2026-01-24 00:24:34 +0000 UTC" firstStartedPulling="2026-01-24 00:24:35.317434301 +0000 UTC m=+1259.347405302" lastFinishedPulling="2026-01-24 00:24:35.732024607 +0000 UTC m=+1259.761995608" observedRunningTime="2026-01-24 00:24:36.338829921 +0000 UTC m=+1260.368800922" watchObservedRunningTime="2026-01-24 00:24:36.396991283 +0000 UTC m=+1260.426962274" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.407963 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d70bd7f-2410-490a-897f-c17149dbc01d" (UID: "0d70bd7f-2410-490a-897f-c17149dbc01d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.446932 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsdvz\" (UniqueName: \"kubernetes.io/projected/0d70bd7f-2410-490a-897f-c17149dbc01d-kube-api-access-fsdvz\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.446969 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.446979 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d70bd7f-2410-490a-897f-c17149dbc01d-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.474508 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0d70bd7f-2410-490a-897f-c17149dbc01d","Type":"ContainerDied","Data":"d5aab805f55f9b9bf75ccb1416d0d46c71a7f8b8d382ea34d87b6fc4caf8ea0a"} Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.474667 4676 scope.go:117] "RemoveContainer" containerID="d5aab805f55f9b9bf75ccb1416d0d46c71a7f8b8d382ea34d87b6fc4caf8ea0a" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.499713 4676 scope.go:117] "RemoveContainer" containerID="9b214cec1541db1b7c8f4c7ce792f15f1a691f1915cb848b2cef3968d5d483b7" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.632622 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.640396 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.662775 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:36 crc kubenswrapper[4676]: E0124 00:24:36.663157 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-api" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.663173 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-api" Jan 24 00:24:36 crc kubenswrapper[4676]: E0124 00:24:36.663189 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-log" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.663195 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-log" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.663396 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-log" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.663412 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" containerName="nova-api-api" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.664285 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.666990 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.681641 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.752667 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.752754 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a8b8762-d78d-44ad-af70-5d53aced9c93-logs\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.753001 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbwbq\" (UniqueName: \"kubernetes.io/projected/3a8b8762-d78d-44ad-af70-5d53aced9c93-kube-api-access-zbwbq\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.753119 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-config-data\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.854184 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.854272 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a8b8762-d78d-44ad-af70-5d53aced9c93-logs\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.854344 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbwbq\" (UniqueName: \"kubernetes.io/projected/3a8b8762-d78d-44ad-af70-5d53aced9c93-kube-api-access-zbwbq\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.854398 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-config-data\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.854810 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a8b8762-d78d-44ad-af70-5d53aced9c93-logs\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.857982 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.862928 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-config-data\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.878934 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbwbq\" (UniqueName: \"kubernetes.io/projected/3a8b8762-d78d-44ad-af70-5d53aced9c93-kube-api-access-zbwbq\") pod \"nova-api-0\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " pod="openstack/nova-api-0" Jan 24 00:24:36 crc kubenswrapper[4676]: I0124 00:24:36.982874 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.084734 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.166776 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-tls-certs\") pod \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.167258 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-scripts\") pod \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.167846 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-secret-key\") pod \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.168584 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-logs\") pod \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.168636 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-config-data\") pod \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.168660 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-combined-ca-bundle\") pod \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.168728 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwffk\" (UniqueName: \"kubernetes.io/projected/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-kube-api-access-cwffk\") pod \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\" (UID: \"9d2451a0-4896-46e4-9b9e-e309ccdf02f2\") " Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.169894 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-logs" (OuterVolumeSpecName: "logs") pod "9d2451a0-4896-46e4-9b9e-e309ccdf02f2" (UID: "9d2451a0-4896-46e4-9b9e-e309ccdf02f2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.174414 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-kube-api-access-cwffk" (OuterVolumeSpecName: "kube-api-access-cwffk") pod "9d2451a0-4896-46e4-9b9e-e309ccdf02f2" (UID: "9d2451a0-4896-46e4-9b9e-e309ccdf02f2"). InnerVolumeSpecName "kube-api-access-cwffk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.177554 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9d2451a0-4896-46e4-9b9e-e309ccdf02f2" (UID: "9d2451a0-4896-46e4-9b9e-e309ccdf02f2"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.206132 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-config-data" (OuterVolumeSpecName: "config-data") pod "9d2451a0-4896-46e4-9b9e-e309ccdf02f2" (UID: "9d2451a0-4896-46e4-9b9e-e309ccdf02f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.208742 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-scripts" (OuterVolumeSpecName: "scripts") pod "9d2451a0-4896-46e4-9b9e-e309ccdf02f2" (UID: "9d2451a0-4896-46e4-9b9e-e309ccdf02f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.220027 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d2451a0-4896-46e4-9b9e-e309ccdf02f2" (UID: "9d2451a0-4896-46e4-9b9e-e309ccdf02f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.270513 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "9d2451a0-4896-46e4-9b9e-e309ccdf02f2" (UID: "9d2451a0-4896-46e4-9b9e-e309ccdf02f2"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.271610 4676 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.271637 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.271646 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.271655 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.271663 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwffk\" (UniqueName: \"kubernetes.io/projected/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-kube-api-access-cwffk\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.271675 4676 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.271684 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2451a0-4896-46e4-9b9e-e309ccdf02f2-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.308099 4676 generic.go:334] "Generic (PLEG): container finished" podID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerID="ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb" exitCode=137 Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.308153 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf988b4bd-ls7hp" event={"ID":"9d2451a0-4896-46e4-9b9e-e309ccdf02f2","Type":"ContainerDied","Data":"ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb"} Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.308172 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-bf988b4bd-ls7hp" event={"ID":"9d2451a0-4896-46e4-9b9e-e309ccdf02f2","Type":"ContainerDied","Data":"e16c7aef4fd28d5aab8b64abddb94def56344a74657a3085b40c7035b723d59e"} Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.308188 4676 scope.go:117] "RemoveContainer" containerID="6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.308268 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-bf988b4bd-ls7hp" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.315606 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"32d678d8-bc5d-4993-afc3-c6700d53b00e","Type":"ContainerStarted","Data":"9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142"} Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.347915 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.347893365 podStartE2EDuration="2.347893365s" podCreationTimestamp="2026-01-24 00:24:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:37.330698431 +0000 UTC m=+1261.360669432" watchObservedRunningTime="2026-01-24 00:24:37.347893365 +0000 UTC m=+1261.377864356" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.419156 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-bf988b4bd-ls7hp"] Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.429073 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-bf988b4bd-ls7hp"] Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.550149 4676 scope.go:117] "RemoveContainer" containerID="ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.567535 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.575758 4676 scope.go:117] "RemoveContainer" containerID="6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64" Jan 24 00:24:37 crc kubenswrapper[4676]: E0124 00:24:37.580017 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64\": container with ID starting with 6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64 not found: ID does not exist" containerID="6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.580056 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64"} err="failed to get container status \"6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64\": rpc error: code = NotFound desc = could not find container \"6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64\": container with ID starting with 6c39f8f7eb482d941eb276b021603978c3c819c52d54e3619ae39a0cd7e29c64 not found: ID does not exist" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.580083 4676 scope.go:117] "RemoveContainer" containerID="ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb" Jan 24 00:24:37 crc kubenswrapper[4676]: E0124 00:24:37.580575 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb\": container with ID starting with ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb not found: ID does not exist" containerID="ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb" Jan 24 00:24:37 crc kubenswrapper[4676]: I0124 00:24:37.580609 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb"} err="failed to get container status \"ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb\": rpc error: code = NotFound desc = could not find container \"ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb\": container with ID starting with ca1454b684ac38a1446a4ef033530125143bc9a4b12de7af96f54377b563b8fb not found: ID does not exist" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.286522 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d70bd7f-2410-490a-897f-c17149dbc01d" path="/var/lib/kubelet/pods/0d70bd7f-2410-490a-897f-c17149dbc01d/volumes" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.304769 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" path="/var/lib/kubelet/pods/9d2451a0-4896-46e4-9b9e-e309ccdf02f2/volumes" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.365474 4676 generic.go:334] "Generic (PLEG): container finished" podID="30206279-8386-4575-9457-e76760505e8d" containerID="74fc38a94fa794a834f28752c4c0e574b0f6bd303f04b96af63769a91e8933fe" exitCode=0 Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.365551 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerDied","Data":"74fc38a94fa794a834f28752c4c0e574b0f6bd303f04b96af63769a91e8933fe"} Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.378722 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a8b8762-d78d-44ad-af70-5d53aced9c93","Type":"ContainerStarted","Data":"261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0"} Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.378756 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a8b8762-d78d-44ad-af70-5d53aced9c93","Type":"ContainerStarted","Data":"3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1"} Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.378767 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a8b8762-d78d-44ad-af70-5d53aced9c93","Type":"ContainerStarted","Data":"000b1ff3c0019bc3bb4c62b73efcef9044b9001595b681004d951af702c62578"} Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.488008 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.508045 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.5080280630000003 podStartE2EDuration="2.508028063s" podCreationTimestamp="2026-01-24 00:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:38.400341642 +0000 UTC m=+1262.430312643" watchObservedRunningTime="2026-01-24 00:24:38.508028063 +0000 UTC m=+1262.537999064" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.610421 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-config-data\") pod \"30206279-8386-4575-9457-e76760505e8d\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.610499 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-sg-core-conf-yaml\") pod \"30206279-8386-4575-9457-e76760505e8d\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.610543 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhdb9\" (UniqueName: \"kubernetes.io/projected/30206279-8386-4575-9457-e76760505e8d-kube-api-access-qhdb9\") pod \"30206279-8386-4575-9457-e76760505e8d\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.610610 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-run-httpd\") pod \"30206279-8386-4575-9457-e76760505e8d\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.610649 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-log-httpd\") pod \"30206279-8386-4575-9457-e76760505e8d\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.610795 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-combined-ca-bundle\") pod \"30206279-8386-4575-9457-e76760505e8d\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.610831 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-scripts\") pod \"30206279-8386-4575-9457-e76760505e8d\" (UID: \"30206279-8386-4575-9457-e76760505e8d\") " Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.611209 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "30206279-8386-4575-9457-e76760505e8d" (UID: "30206279-8386-4575-9457-e76760505e8d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.611359 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "30206279-8386-4575-9457-e76760505e8d" (UID: "30206279-8386-4575-9457-e76760505e8d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.611773 4676 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.611788 4676 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/30206279-8386-4575-9457-e76760505e8d-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.615484 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30206279-8386-4575-9457-e76760505e8d-kube-api-access-qhdb9" (OuterVolumeSpecName: "kube-api-access-qhdb9") pod "30206279-8386-4575-9457-e76760505e8d" (UID: "30206279-8386-4575-9457-e76760505e8d"). InnerVolumeSpecName "kube-api-access-qhdb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.619868 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-scripts" (OuterVolumeSpecName: "scripts") pod "30206279-8386-4575-9457-e76760505e8d" (UID: "30206279-8386-4575-9457-e76760505e8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.648795 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "30206279-8386-4575-9457-e76760505e8d" (UID: "30206279-8386-4575-9457-e76760505e8d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.708940 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30206279-8386-4575-9457-e76760505e8d" (UID: "30206279-8386-4575-9457-e76760505e8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.713655 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.713760 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.713851 4676 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.713918 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhdb9\" (UniqueName: \"kubernetes.io/projected/30206279-8386-4575-9457-e76760505e8d-kube-api-access-qhdb9\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.733530 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-config-data" (OuterVolumeSpecName: "config-data") pod "30206279-8386-4575-9457-e76760505e8d" (UID: "30206279-8386-4575-9457-e76760505e8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:38 crc kubenswrapper[4676]: I0124 00:24:38.815457 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30206279-8386-4575-9457-e76760505e8d-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.388113 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.391040 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"30206279-8386-4575-9457-e76760505e8d","Type":"ContainerDied","Data":"27bde3b4f936759dfdde3e8fe3259a873617d7354ea2e22a706cd0edd462041d"} Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.391099 4676 scope.go:117] "RemoveContainer" containerID="cf2f47547ef5fe4c3cd5d6009235ea3a947829e0913a6e11bb70499604aaa440" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.425910 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.428171 4676 scope.go:117] "RemoveContainer" containerID="993ed8d33422939d102af22915be4b96f85bac159232e43af92e15ee1383a46a" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.436643 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.446096 4676 scope.go:117] "RemoveContainer" containerID="74fc38a94fa794a834f28752c4c0e574b0f6bd303f04b96af63769a91e8933fe" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.465152 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:24:39 crc kubenswrapper[4676]: E0124 00:24:39.465599 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.465621 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" Jan 24 00:24:39 crc kubenswrapper[4676]: E0124 00:24:39.465633 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="ceilometer-central-agent" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.465644 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="ceilometer-central-agent" Jan 24 00:24:39 crc kubenswrapper[4676]: E0124 00:24:39.465673 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="proxy-httpd" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.465682 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="proxy-httpd" Jan 24 00:24:39 crc kubenswrapper[4676]: E0124 00:24:39.465696 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="ceilometer-notification-agent" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.465704 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="ceilometer-notification-agent" Jan 24 00:24:39 crc kubenswrapper[4676]: E0124 00:24:39.465716 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon-log" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.469724 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon-log" Jan 24 00:24:39 crc kubenswrapper[4676]: E0124 00:24:39.469747 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.469755 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" Jan 24 00:24:39 crc kubenswrapper[4676]: E0124 00:24:39.469769 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="sg-core" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.469777 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="sg-core" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.470084 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon-log" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.470116 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.470128 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="sg-core" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.470143 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="proxy-httpd" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.470156 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="ceilometer-central-agent" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.470173 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="30206279-8386-4575-9457-e76760505e8d" containerName="ceilometer-notification-agent" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.470634 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d2451a0-4896-46e4-9b9e-e309ccdf02f2" containerName="horizon" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.470795 4676 scope.go:117] "RemoveContainer" containerID="acf37fd15df58f0c480e8ca3019535bc0893909734220600f3f004d89aa3edc9" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.476042 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.476599 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.481965 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.481985 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.482247 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.528859 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.528921 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-config-data\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.528951 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-log-httpd\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.529020 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.529055 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-scripts\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.529169 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq2lz\" (UniqueName: \"kubernetes.io/projected/3c825a1a-2a44-4f85-9271-0614aa8a07a4-kube-api-access-nq2lz\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.529194 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.529211 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-run-httpd\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.630629 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.630686 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-scripts\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.630765 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq2lz\" (UniqueName: \"kubernetes.io/projected/3c825a1a-2a44-4f85-9271-0614aa8a07a4-kube-api-access-nq2lz\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.630789 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.630809 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-run-httpd\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.630830 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.630852 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-config-data\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.630873 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-log-httpd\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.631357 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-log-httpd\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.631485 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-run-httpd\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.635116 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-scripts\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.636621 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-config-data\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.637446 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.637947 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.639777 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.663314 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq2lz\" (UniqueName: \"kubernetes.io/projected/3c825a1a-2a44-4f85-9271-0614aa8a07a4-kube-api-access-nq2lz\") pod \"ceilometer-0\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " pod="openstack/ceilometer-0" Jan 24 00:24:39 crc kubenswrapper[4676]: I0124 00:24:39.810696 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:24:40 crc kubenswrapper[4676]: I0124 00:24:40.270562 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30206279-8386-4575-9457-e76760505e8d" path="/var/lib/kubelet/pods/30206279-8386-4575-9457-e76760505e8d/volumes" Jan 24 00:24:40 crc kubenswrapper[4676]: I0124 00:24:40.383143 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:24:40 crc kubenswrapper[4676]: I0124 00:24:40.647088 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 24 00:24:40 crc kubenswrapper[4676]: I0124 00:24:40.733344 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Jan 24 00:24:41 crc kubenswrapper[4676]: I0124 00:24:41.413227 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerStarted","Data":"779dd89d442e8cfef9c213f780a757ec699bc0eef79a17d12ff2edf42b44250b"} Jan 24 00:24:41 crc kubenswrapper[4676]: I0124 00:24:41.413600 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerStarted","Data":"9944d0b2b4ee648d3da07eb6834b39006236c631ca51eb113778101cb554219b"} Jan 24 00:24:42 crc kubenswrapper[4676]: I0124 00:24:42.423488 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerStarted","Data":"33f9fd41388e48671ab7c03de55b7972b9dfd2592bc4af417c289b099fe20556"} Jan 24 00:24:43 crc kubenswrapper[4676]: I0124 00:24:43.435403 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerStarted","Data":"e7c4de9128cd6e6e3de76a1907fc12bdd1ffdc61da41090dc5b038476ebe4f1a"} Jan 24 00:24:44 crc kubenswrapper[4676]: I0124 00:24:44.444450 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerStarted","Data":"99145c476bd92edc34903b04d49544d24801d588d3c6e0a3159018baae1cfdbf"} Jan 24 00:24:44 crc kubenswrapper[4676]: I0124 00:24:44.445929 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 24 00:24:44 crc kubenswrapper[4676]: I0124 00:24:44.471143 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8863444390000002 podStartE2EDuration="5.471125986s" podCreationTimestamp="2026-01-24 00:24:39 +0000 UTC" firstStartedPulling="2026-01-24 00:24:40.392638072 +0000 UTC m=+1264.422609093" lastFinishedPulling="2026-01-24 00:24:43.977419629 +0000 UTC m=+1268.007390640" observedRunningTime="2026-01-24 00:24:44.470003182 +0000 UTC m=+1268.499974193" watchObservedRunningTime="2026-01-24 00:24:44.471125986 +0000 UTC m=+1268.501096977" Jan 24 00:24:44 crc kubenswrapper[4676]: I0124 00:24:44.624233 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 24 00:24:45 crc kubenswrapper[4676]: I0124 00:24:45.647343 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 24 00:24:45 crc kubenswrapper[4676]: I0124 00:24:45.697336 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 24 00:24:46 crc kubenswrapper[4676]: I0124 00:24:46.505350 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 24 00:24:46 crc kubenswrapper[4676]: I0124 00:24:46.986090 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 24 00:24:46 crc kubenswrapper[4676]: I0124 00:24:46.986149 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 24 00:24:48 crc kubenswrapper[4676]: I0124 00:24:48.069752 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 24 00:24:48 crc kubenswrapper[4676]: I0124 00:24:48.069848 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 24 00:24:53 crc kubenswrapper[4676]: E0124 00:24:53.385073 4676 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84633021_2bad_4066_a01c_44fef6902524.slice/crio-conmon-6301d47d4294644c70fd618b97e56c019270427fbdab822047ac3bfe88e7e594.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84633021_2bad_4066_a01c_44fef6902524.slice/crio-6301d47d4294644c70fd618b97e56c019270427fbdab822047ac3bfe88e7e594.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65324a0f_f62a_4807_85d2_e4b607b2a0b4.slice/crio-e4451937240be2f867c73971e845e8f6b7d4b6bee35057ae40b7f593e4b5d73c.scope\": RecentStats: unable to find data in memory cache]" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.534715 4676 generic.go:334] "Generic (PLEG): container finished" podID="65324a0f-f62a-4807-85d2-e4b607b2a0b4" containerID="e4451937240be2f867c73971e845e8f6b7d4b6bee35057ae40b7f593e4b5d73c" exitCode=137 Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.534806 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"65324a0f-f62a-4807-85d2-e4b607b2a0b4","Type":"ContainerDied","Data":"e4451937240be2f867c73971e845e8f6b7d4b6bee35057ae40b7f593e4b5d73c"} Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.537035 4676 generic.go:334] "Generic (PLEG): container finished" podID="84633021-2bad-4066-a01c-44fef6902524" containerID="6301d47d4294644c70fd618b97e56c019270427fbdab822047ac3bfe88e7e594" exitCode=137 Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.537161 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84633021-2bad-4066-a01c-44fef6902524","Type":"ContainerDied","Data":"6301d47d4294644c70fd618b97e56c019270427fbdab822047ac3bfe88e7e594"} Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.537262 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"84633021-2bad-4066-a01c-44fef6902524","Type":"ContainerDied","Data":"6e82aa991fb418472d65f2acfb31fecbeef7f12dbcdb1246d40b5110c0b6ab65"} Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.537344 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e82aa991fb418472d65f2acfb31fecbeef7f12dbcdb1246d40b5110c0b6ab65" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.642789 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.650774 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.809085 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9hf5\" (UniqueName: \"kubernetes.io/projected/65324a0f-f62a-4807-85d2-e4b607b2a0b4-kube-api-access-p9hf5\") pod \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.810326 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-config-data\") pod \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.810492 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84633021-2bad-4066-a01c-44fef6902524-logs\") pod \"84633021-2bad-4066-a01c-44fef6902524\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.810948 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f77n\" (UniqueName: \"kubernetes.io/projected/84633021-2bad-4066-a01c-44fef6902524-kube-api-access-8f77n\") pod \"84633021-2bad-4066-a01c-44fef6902524\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.811604 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-combined-ca-bundle\") pod \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\" (UID: \"65324a0f-f62a-4807-85d2-e4b607b2a0b4\") " Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.811748 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-combined-ca-bundle\") pod \"84633021-2bad-4066-a01c-44fef6902524\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.811895 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-config-data\") pod \"84633021-2bad-4066-a01c-44fef6902524\" (UID: \"84633021-2bad-4066-a01c-44fef6902524\") " Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.811139 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84633021-2bad-4066-a01c-44fef6902524-logs" (OuterVolumeSpecName: "logs") pod "84633021-2bad-4066-a01c-44fef6902524" (UID: "84633021-2bad-4066-a01c-44fef6902524"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.812891 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/84633021-2bad-4066-a01c-44fef6902524-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.814875 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65324a0f-f62a-4807-85d2-e4b607b2a0b4-kube-api-access-p9hf5" (OuterVolumeSpecName: "kube-api-access-p9hf5") pod "65324a0f-f62a-4807-85d2-e4b607b2a0b4" (UID: "65324a0f-f62a-4807-85d2-e4b607b2a0b4"). InnerVolumeSpecName "kube-api-access-p9hf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.816911 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84633021-2bad-4066-a01c-44fef6902524-kube-api-access-8f77n" (OuterVolumeSpecName: "kube-api-access-8f77n") pod "84633021-2bad-4066-a01c-44fef6902524" (UID: "84633021-2bad-4066-a01c-44fef6902524"). InnerVolumeSpecName "kube-api-access-8f77n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.839440 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-config-data" (OuterVolumeSpecName: "config-data") pod "84633021-2bad-4066-a01c-44fef6902524" (UID: "84633021-2bad-4066-a01c-44fef6902524"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.842254 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-config-data" (OuterVolumeSpecName: "config-data") pod "65324a0f-f62a-4807-85d2-e4b607b2a0b4" (UID: "65324a0f-f62a-4807-85d2-e4b607b2a0b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.847522 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "84633021-2bad-4066-a01c-44fef6902524" (UID: "84633021-2bad-4066-a01c-44fef6902524"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.852457 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65324a0f-f62a-4807-85d2-e4b607b2a0b4" (UID: "65324a0f-f62a-4807-85d2-e4b607b2a0b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.915352 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.915408 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9hf5\" (UniqueName: \"kubernetes.io/projected/65324a0f-f62a-4807-85d2-e4b607b2a0b4-kube-api-access-p9hf5\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.915426 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.915438 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f77n\" (UniqueName: \"kubernetes.io/projected/84633021-2bad-4066-a01c-44fef6902524-kube-api-access-8f77n\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.915450 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65324a0f-f62a-4807-85d2-e4b607b2a0b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:53 crc kubenswrapper[4676]: I0124 00:24:53.915464 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/84633021-2bad-4066-a01c-44fef6902524-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.549197 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"65324a0f-f62a-4807-85d2-e4b607b2a0b4","Type":"ContainerDied","Data":"9d99431ded08302b9fd0007acf0016950fd97f2998e3a74b9630c85127cb16f9"} Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.549552 4676 scope.go:117] "RemoveContainer" containerID="e4451937240be2f867c73971e845e8f6b7d4b6bee35057ae40b7f593e4b5d73c" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.549242 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.550986 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.584224 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.602590 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.623433 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.632425 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.652104 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 24 00:24:54 crc kubenswrapper[4676]: E0124 00:24:54.652609 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84633021-2bad-4066-a01c-44fef6902524" containerName="nova-metadata-log" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.652627 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="84633021-2bad-4066-a01c-44fef6902524" containerName="nova-metadata-log" Jan 24 00:24:54 crc kubenswrapper[4676]: E0124 00:24:54.652640 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84633021-2bad-4066-a01c-44fef6902524" containerName="nova-metadata-metadata" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.652648 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="84633021-2bad-4066-a01c-44fef6902524" containerName="nova-metadata-metadata" Jan 24 00:24:54 crc kubenswrapper[4676]: E0124 00:24:54.652668 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65324a0f-f62a-4807-85d2-e4b607b2a0b4" containerName="nova-cell1-novncproxy-novncproxy" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.652680 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="65324a0f-f62a-4807-85d2-e4b607b2a0b4" containerName="nova-cell1-novncproxy-novncproxy" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.652906 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="84633021-2bad-4066-a01c-44fef6902524" containerName="nova-metadata-metadata" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.652935 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="65324a0f-f62a-4807-85d2-e4b607b2a0b4" containerName="nova-cell1-novncproxy-novncproxy" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.652945 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="84633021-2bad-4066-a01c-44fef6902524" containerName="nova-metadata-log" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.653653 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.665875 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.666113 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.666353 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.678507 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.685247 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.687145 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.691474 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.691677 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.695746 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.732940 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.733006 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6784012-a1a0-4647-929f-50788a8dc5bb-logs\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.733051 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.733075 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.733118 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.733152 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-config-data\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.733185 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.733209 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prt8v\" (UniqueName: \"kubernetes.io/projected/b1499ed8-1041-4980-b9da-cb957cbf215c-kube-api-access-prt8v\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.733231 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.733272 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4dfw\" (UniqueName: \"kubernetes.io/projected/d6784012-a1a0-4647-929f-50788a8dc5bb-kube-api-access-p4dfw\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834279 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4dfw\" (UniqueName: \"kubernetes.io/projected/d6784012-a1a0-4647-929f-50788a8dc5bb-kube-api-access-p4dfw\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834340 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834393 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6784012-a1a0-4647-929f-50788a8dc5bb-logs\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834434 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834454 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834491 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834519 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-config-data\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834542 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834560 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prt8v\" (UniqueName: \"kubernetes.io/projected/b1499ed8-1041-4980-b9da-cb957cbf215c-kube-api-access-prt8v\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.834580 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.835279 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6784012-a1a0-4647-929f-50788a8dc5bb-logs\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.838941 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.841459 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-config-data\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.842866 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.846760 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.846841 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.847897 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.851036 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1499ed8-1041-4980-b9da-cb957cbf215c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.853010 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4dfw\" (UniqueName: \"kubernetes.io/projected/d6784012-a1a0-4647-929f-50788a8dc5bb-kube-api-access-p4dfw\") pod \"nova-metadata-0\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " pod="openstack/nova-metadata-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.860624 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prt8v\" (UniqueName: \"kubernetes.io/projected/b1499ed8-1041-4980-b9da-cb957cbf215c-kube-api-access-prt8v\") pod \"nova-cell1-novncproxy-0\" (UID: \"b1499ed8-1041-4980-b9da-cb957cbf215c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:54 crc kubenswrapper[4676]: I0124 00:24:54.983846 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:24:55 crc kubenswrapper[4676]: I0124 00:24:55.026322 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:24:55 crc kubenswrapper[4676]: I0124 00:24:55.519864 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 24 00:24:55 crc kubenswrapper[4676]: W0124 00:24:55.530642 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1499ed8_1041_4980_b9da_cb957cbf215c.slice/crio-6c92d02f4ee03a49f06a51e0afb40d6b241270cd7d28e0e7dfe00dcde1fa435f WatchSource:0}: Error finding container 6c92d02f4ee03a49f06a51e0afb40d6b241270cd7d28e0e7dfe00dcde1fa435f: Status 404 returned error can't find the container with id 6c92d02f4ee03a49f06a51e0afb40d6b241270cd7d28e0e7dfe00dcde1fa435f Jan 24 00:24:55 crc kubenswrapper[4676]: I0124 00:24:55.539344 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:24:55 crc kubenswrapper[4676]: I0124 00:24:55.562257 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b1499ed8-1041-4980-b9da-cb957cbf215c","Type":"ContainerStarted","Data":"6c92d02f4ee03a49f06a51e0afb40d6b241270cd7d28e0e7dfe00dcde1fa435f"} Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.267890 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65324a0f-f62a-4807-85d2-e4b607b2a0b4" path="/var/lib/kubelet/pods/65324a0f-f62a-4807-85d2-e4b607b2a0b4/volumes" Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.268962 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84633021-2bad-4066-a01c-44fef6902524" path="/var/lib/kubelet/pods/84633021-2bad-4066-a01c-44fef6902524/volumes" Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.578333 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b1499ed8-1041-4980-b9da-cb957cbf215c","Type":"ContainerStarted","Data":"bbe545070675448028e845841b263321ae62b0b7655e0707a2be07d263d064e6"} Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.583698 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6784012-a1a0-4647-929f-50788a8dc5bb","Type":"ContainerStarted","Data":"069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b"} Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.583741 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6784012-a1a0-4647-929f-50788a8dc5bb","Type":"ContainerStarted","Data":"dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1"} Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.583751 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6784012-a1a0-4647-929f-50788a8dc5bb","Type":"ContainerStarted","Data":"01fe7e14f549bff2abe35a4d5403466f7bbd36dfe57daa6ef4e421db5dcd7cf7"} Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.605069 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.605042322 podStartE2EDuration="2.605042322s" podCreationTimestamp="2026-01-24 00:24:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:56.601279567 +0000 UTC m=+1280.631250598" watchObservedRunningTime="2026-01-24 00:24:56.605042322 +0000 UTC m=+1280.635013363" Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.635431 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.635408118 podStartE2EDuration="2.635408118s" podCreationTimestamp="2026-01-24 00:24:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:56.631283852 +0000 UTC m=+1280.661254853" watchObservedRunningTime="2026-01-24 00:24:56.635408118 +0000 UTC m=+1280.665379129" Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.988940 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.989355 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.989689 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.989815 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 24 00:24:56 crc kubenswrapper[4676]: I0124 00:24:56.998729 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.003554 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.178249 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-tkk74"] Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.179639 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.200087 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-tkk74"] Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.278999 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.279172 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.279196 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.279227 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-config\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.279243 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.279326 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5hbp\" (UniqueName: \"kubernetes.io/projected/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-kube-api-access-s5hbp\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.381182 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.381305 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.381328 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.381348 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-config\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.381366 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.381423 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5hbp\" (UniqueName: \"kubernetes.io/projected/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-kube-api-access-s5hbp\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.382390 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.382430 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.382466 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.383390 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.383725 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-config\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.402403 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5hbp\" (UniqueName: \"kubernetes.io/projected/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-kube-api-access-s5hbp\") pod \"dnsmasq-dns-cd5cbd7b9-tkk74\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:57 crc kubenswrapper[4676]: I0124 00:24:57.496844 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:24:58 crc kubenswrapper[4676]: I0124 00:24:58.231222 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-tkk74"] Jan 24 00:24:58 crc kubenswrapper[4676]: W0124 00:24:58.234054 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a3231ad_bc2c_4a51_813d_cfd1a11c4fc4.slice/crio-85858d9bee0a25807ae1aa70febaf41556800347d26ef61522b473ffb2662a4f WatchSource:0}: Error finding container 85858d9bee0a25807ae1aa70febaf41556800347d26ef61522b473ffb2662a4f: Status 404 returned error can't find the container with id 85858d9bee0a25807ae1aa70febaf41556800347d26ef61522b473ffb2662a4f Jan 24 00:24:58 crc kubenswrapper[4676]: I0124 00:24:58.619675 4676 generic.go:334] "Generic (PLEG): container finished" podID="6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" containerID="e9acf51095b7dd6a0ad30ae069a3cb2ce0ad324415a9a8bc556468f57ab1c34d" exitCode=0 Jan 24 00:24:58 crc kubenswrapper[4676]: I0124 00:24:58.619779 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" event={"ID":"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4","Type":"ContainerDied","Data":"e9acf51095b7dd6a0ad30ae069a3cb2ce0ad324415a9a8bc556468f57ab1c34d"} Jan 24 00:24:58 crc kubenswrapper[4676]: I0124 00:24:58.619916 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" event={"ID":"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4","Type":"ContainerStarted","Data":"85858d9bee0a25807ae1aa70febaf41556800347d26ef61522b473ffb2662a4f"} Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.518598 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.628501 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" event={"ID":"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4","Type":"ContainerStarted","Data":"8189f6b4420a2e8c611c03a3298dcf14f84845a898137968cf5f45e6c0c37685"} Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.628622 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-log" containerID="cri-o://3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1" gracePeriod=30 Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.629119 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-api" containerID="cri-o://261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0" gracePeriod=30 Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.661992 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" podStartSLOduration=2.66197296 podStartE2EDuration="2.66197296s" podCreationTimestamp="2026-01-24 00:24:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:24:59.65343571 +0000 UTC m=+1283.683406711" watchObservedRunningTime="2026-01-24 00:24:59.66197296 +0000 UTC m=+1283.691943961" Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.888983 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.889644 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="ceilometer-central-agent" containerID="cri-o://779dd89d442e8cfef9c213f780a757ec699bc0eef79a17d12ff2edf42b44250b" gracePeriod=30 Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.890489 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="proxy-httpd" containerID="cri-o://99145c476bd92edc34903b04d49544d24801d588d3c6e0a3159018baae1cfdbf" gracePeriod=30 Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.890559 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="sg-core" containerID="cri-o://e7c4de9128cd6e6e3de76a1907fc12bdd1ffdc61da41090dc5b038476ebe4f1a" gracePeriod=30 Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.890606 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="ceilometer-notification-agent" containerID="cri-o://33f9fd41388e48671ab7c03de55b7972b9dfd2592bc4af417c289b099fe20556" gracePeriod=30 Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.912732 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.193:3000/\": EOF" Jan 24 00:24:59 crc kubenswrapper[4676]: I0124 00:24:59.984788 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.026925 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.027124 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.638947 4676 generic.go:334] "Generic (PLEG): container finished" podID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerID="99145c476bd92edc34903b04d49544d24801d588d3c6e0a3159018baae1cfdbf" exitCode=0 Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.639939 4676 generic.go:334] "Generic (PLEG): container finished" podID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerID="e7c4de9128cd6e6e3de76a1907fc12bdd1ffdc61da41090dc5b038476ebe4f1a" exitCode=2 Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.638987 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerDied","Data":"99145c476bd92edc34903b04d49544d24801d588d3c6e0a3159018baae1cfdbf"} Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.640042 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerDied","Data":"e7c4de9128cd6e6e3de76a1907fc12bdd1ffdc61da41090dc5b038476ebe4f1a"} Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.640057 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerDied","Data":"779dd89d442e8cfef9c213f780a757ec699bc0eef79a17d12ff2edf42b44250b"} Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.640015 4676 generic.go:334] "Generic (PLEG): container finished" podID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerID="779dd89d442e8cfef9c213f780a757ec699bc0eef79a17d12ff2edf42b44250b" exitCode=0 Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.641862 4676 generic.go:334] "Generic (PLEG): container finished" podID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerID="3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1" exitCode=143 Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.642020 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a8b8762-d78d-44ad-af70-5d53aced9c93","Type":"ContainerDied","Data":"3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1"} Jan 24 00:25:00 crc kubenswrapper[4676]: I0124 00:25:00.642138 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:25:02 crc kubenswrapper[4676]: I0124 00:25:02.697355 4676 generic.go:334] "Generic (PLEG): container finished" podID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerID="33f9fd41388e48671ab7c03de55b7972b9dfd2592bc4af417c289b099fe20556" exitCode=0 Jan 24 00:25:02 crc kubenswrapper[4676]: I0124 00:25:02.697473 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerDied","Data":"33f9fd41388e48671ab7c03de55b7972b9dfd2592bc4af417c289b099fe20556"} Jan 24 00:25:02 crc kubenswrapper[4676]: I0124 00:25:02.975493 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.096536 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-config-data\") pod \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.097083 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq2lz\" (UniqueName: \"kubernetes.io/projected/3c825a1a-2a44-4f85-9271-0614aa8a07a4-kube-api-access-nq2lz\") pod \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.097149 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-log-httpd\") pod \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.097220 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-scripts\") pod \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.097271 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-run-httpd\") pod \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.097770 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3c825a1a-2a44-4f85-9271-0614aa8a07a4" (UID: "3c825a1a-2a44-4f85-9271-0614aa8a07a4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.098368 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3c825a1a-2a44-4f85-9271-0614aa8a07a4" (UID: "3c825a1a-2a44-4f85-9271-0614aa8a07a4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.098522 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-combined-ca-bundle\") pod \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.098575 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-sg-core-conf-yaml\") pod \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.098642 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-ceilometer-tls-certs\") pod \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\" (UID: \"3c825a1a-2a44-4f85-9271-0614aa8a07a4\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.099644 4676 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.099671 4676 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c825a1a-2a44-4f85-9271-0614aa8a07a4-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.122799 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-scripts" (OuterVolumeSpecName: "scripts") pod "3c825a1a-2a44-4f85-9271-0614aa8a07a4" (UID: "3c825a1a-2a44-4f85-9271-0614aa8a07a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.128095 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c825a1a-2a44-4f85-9271-0614aa8a07a4-kube-api-access-nq2lz" (OuterVolumeSpecName: "kube-api-access-nq2lz") pod "3c825a1a-2a44-4f85-9271-0614aa8a07a4" (UID: "3c825a1a-2a44-4f85-9271-0614aa8a07a4"). InnerVolumeSpecName "kube-api-access-nq2lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.182428 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3c825a1a-2a44-4f85-9271-0614aa8a07a4" (UID: "3c825a1a-2a44-4f85-9271-0614aa8a07a4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.202833 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq2lz\" (UniqueName: \"kubernetes.io/projected/3c825a1a-2a44-4f85-9271-0614aa8a07a4-kube-api-access-nq2lz\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.202862 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.202872 4676 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.211043 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3c825a1a-2a44-4f85-9271-0614aa8a07a4" (UID: "3c825a1a-2a44-4f85-9271-0614aa8a07a4"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.231007 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c825a1a-2a44-4f85-9271-0614aa8a07a4" (UID: "3c825a1a-2a44-4f85-9271-0614aa8a07a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.243571 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.264935 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-config-data" (OuterVolumeSpecName: "config-data") pod "3c825a1a-2a44-4f85-9271-0614aa8a07a4" (UID: "3c825a1a-2a44-4f85-9271-0614aa8a07a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.307762 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbwbq\" (UniqueName: \"kubernetes.io/projected/3a8b8762-d78d-44ad-af70-5d53aced9c93-kube-api-access-zbwbq\") pod \"3a8b8762-d78d-44ad-af70-5d53aced9c93\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.308476 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a8b8762-d78d-44ad-af70-5d53aced9c93-logs\") pod \"3a8b8762-d78d-44ad-af70-5d53aced9c93\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.308527 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-config-data\") pod \"3a8b8762-d78d-44ad-af70-5d53aced9c93\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.308701 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-combined-ca-bundle\") pod \"3a8b8762-d78d-44ad-af70-5d53aced9c93\" (UID: \"3a8b8762-d78d-44ad-af70-5d53aced9c93\") " Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.308861 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a8b8762-d78d-44ad-af70-5d53aced9c93-logs" (OuterVolumeSpecName: "logs") pod "3a8b8762-d78d-44ad-af70-5d53aced9c93" (UID: "3a8b8762-d78d-44ad-af70-5d53aced9c93"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.309702 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a8b8762-d78d-44ad-af70-5d53aced9c93-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.309717 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.309726 4676 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.309746 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c825a1a-2a44-4f85-9271-0614aa8a07a4-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.329436 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a8b8762-d78d-44ad-af70-5d53aced9c93-kube-api-access-zbwbq" (OuterVolumeSpecName: "kube-api-access-zbwbq") pod "3a8b8762-d78d-44ad-af70-5d53aced9c93" (UID: "3a8b8762-d78d-44ad-af70-5d53aced9c93"). InnerVolumeSpecName "kube-api-access-zbwbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.334392 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a8b8762-d78d-44ad-af70-5d53aced9c93" (UID: "3a8b8762-d78d-44ad-af70-5d53aced9c93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.348215 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-config-data" (OuterVolumeSpecName: "config-data") pod "3a8b8762-d78d-44ad-af70-5d53aced9c93" (UID: "3a8b8762-d78d-44ad-af70-5d53aced9c93"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.411293 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.411325 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbwbq\" (UniqueName: \"kubernetes.io/projected/3a8b8762-d78d-44ad-af70-5d53aced9c93-kube-api-access-zbwbq\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.411335 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a8b8762-d78d-44ad-af70-5d53aced9c93-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.710761 4676 generic.go:334] "Generic (PLEG): container finished" podID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerID="261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0" exitCode=0 Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.710836 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a8b8762-d78d-44ad-af70-5d53aced9c93","Type":"ContainerDied","Data":"261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0"} Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.710866 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3a8b8762-d78d-44ad-af70-5d53aced9c93","Type":"ContainerDied","Data":"000b1ff3c0019bc3bb4c62b73efcef9044b9001595b681004d951af702c62578"} Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.710886 4676 scope.go:117] "RemoveContainer" containerID="261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.710885 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.716463 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c825a1a-2a44-4f85-9271-0614aa8a07a4","Type":"ContainerDied","Data":"9944d0b2b4ee648d3da07eb6834b39006236c631ca51eb113778101cb554219b"} Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.716552 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.751477 4676 scope.go:117] "RemoveContainer" containerID="3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.755073 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.766593 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.790437 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:03 crc kubenswrapper[4676]: E0124 00:25:03.790875 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-api" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.790897 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-api" Jan 24 00:25:03 crc kubenswrapper[4676]: E0124 00:25:03.790914 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-log" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.790922 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-log" Jan 24 00:25:03 crc kubenswrapper[4676]: E0124 00:25:03.790938 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="ceilometer-central-agent" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.790946 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="ceilometer-central-agent" Jan 24 00:25:03 crc kubenswrapper[4676]: E0124 00:25:03.790960 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="ceilometer-notification-agent" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.790969 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="ceilometer-notification-agent" Jan 24 00:25:03 crc kubenswrapper[4676]: E0124 00:25:03.790994 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="sg-core" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.791002 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="sg-core" Jan 24 00:25:03 crc kubenswrapper[4676]: E0124 00:25:03.791023 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="proxy-httpd" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.791032 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="proxy-httpd" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.791235 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="ceilometer-notification-agent" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.791254 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="sg-core" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.791268 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-log" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.791282 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" containerName="nova-api-api" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.791299 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="ceilometer-central-agent" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.791311 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" containerName="proxy-httpd" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.792463 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.793138 4676 scope.go:117] "RemoveContainer" containerID="261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.796172 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.796340 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 24 00:25:03 crc kubenswrapper[4676]: E0124 00:25:03.796554 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0\": container with ID starting with 261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0 not found: ID does not exist" containerID="261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.796598 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0"} err="failed to get container status \"261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0\": rpc error: code = NotFound desc = could not find container \"261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0\": container with ID starting with 261a0919f777a0b18febac5db55c7874e12a7dc53a2edb25ac33fa7937a21ad0 not found: ID does not exist" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.796627 4676 scope.go:117] "RemoveContainer" containerID="3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1" Jan 24 00:25:03 crc kubenswrapper[4676]: E0124 00:25:03.797909 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1\": container with ID starting with 3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1 not found: ID does not exist" containerID="3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.797944 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1"} err="failed to get container status \"3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1\": rpc error: code = NotFound desc = could not find container \"3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1\": container with ID starting with 3d1351cf150d3d0db98e68d8cbd4da0fc2fc3e8bd6a7873c16725d321e9a9ab1 not found: ID does not exist" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.797971 4676 scope.go:117] "RemoveContainer" containerID="99145c476bd92edc34903b04d49544d24801d588d3c6e0a3159018baae1cfdbf" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.803913 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.805521 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.815009 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.821609 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-public-tls-certs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.821653 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-logs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.821744 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.821768 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-config-data\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.821820 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdbxs\" (UniqueName: \"kubernetes.io/projected/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-kube-api-access-cdbxs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.821912 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.829251 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.831779 4676 scope.go:117] "RemoveContainer" containerID="e7c4de9128cd6e6e3de76a1907fc12bdd1ffdc61da41090dc5b038476ebe4f1a" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.860990 4676 scope.go:117] "RemoveContainer" containerID="33f9fd41388e48671ab7c03de55b7972b9dfd2592bc4af417c289b099fe20556" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.873974 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.876662 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.894134 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.894214 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.894349 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.913072 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.923632 4676 scope.go:117] "RemoveContainer" containerID="779dd89d442e8cfef9c213f780a757ec699bc0eef79a17d12ff2edf42b44250b" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.924635 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.924670 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-public-tls-certs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.924695 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-logs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.924752 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.924771 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-config-data\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.924805 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdbxs\" (UniqueName: \"kubernetes.io/projected/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-kube-api-access-cdbxs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.928579 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-logs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.933353 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.943326 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdbxs\" (UniqueName: \"kubernetes.io/projected/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-kube-api-access-cdbxs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.943947 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-config-data\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.952328 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-public-tls-certs\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:03 crc kubenswrapper[4676]: I0124 00:25:03.956009 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " pod="openstack/nova-api-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.039843 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-scripts\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.039942 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.039981 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58bbd77a-9518-4037-b96b-a1490082fb04-run-httpd\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.040122 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh7ks\" (UniqueName: \"kubernetes.io/projected/58bbd77a-9518-4037-b96b-a1490082fb04-kube-api-access-dh7ks\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.040329 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.040367 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58bbd77a-9518-4037-b96b-a1490082fb04-log-httpd\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.040405 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-config-data\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.040436 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.127197 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.141943 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.142153 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58bbd77a-9518-4037-b96b-a1490082fb04-log-httpd\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.142233 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-config-data\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.142330 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.142448 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-scripts\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.142540 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.142614 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58bbd77a-9518-4037-b96b-a1490082fb04-run-httpd\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.142697 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58bbd77a-9518-4037-b96b-a1490082fb04-log-httpd\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.142787 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh7ks\" (UniqueName: \"kubernetes.io/projected/58bbd77a-9518-4037-b96b-a1490082fb04-kube-api-access-dh7ks\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.143780 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/58bbd77a-9518-4037-b96b-a1490082fb04-run-httpd\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.146291 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-config-data\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.146453 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.147930 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-scripts\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.149892 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.152930 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/58bbd77a-9518-4037-b96b-a1490082fb04-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.174304 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh7ks\" (UniqueName: \"kubernetes.io/projected/58bbd77a-9518-4037-b96b-a1490082fb04-kube-api-access-dh7ks\") pod \"ceilometer-0\" (UID: \"58bbd77a-9518-4037-b96b-a1490082fb04\") " pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.198968 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.309760 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a8b8762-d78d-44ad-af70-5d53aced9c93" path="/var/lib/kubelet/pods/3a8b8762-d78d-44ad-af70-5d53aced9c93/volumes" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.314686 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c825a1a-2a44-4f85-9271-0614aa8a07a4" path="/var/lib/kubelet/pods/3c825a1a-2a44-4f85-9271-0614aa8a07a4/volumes" Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.631890 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.727755 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace","Type":"ContainerStarted","Data":"21b2d30f9d7e2c4a549d8c87dd63e0002ace80a317cb0d0205c94126fe98f6c1"} Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.808657 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 24 00:25:04 crc kubenswrapper[4676]: I0124 00:25:04.985115 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.007906 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.027243 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.027446 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.736804 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace","Type":"ContainerStarted","Data":"01fc81c9856cfd2ef13df58946b52afdb5cba674fdb4423a36d9a11f6c166109"} Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.737048 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace","Type":"ContainerStarted","Data":"71c7f3216e344862d8255e5daeef618a41a9f2fd2e7332ca45fe189abdb2cd5c"} Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.740426 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58bbd77a-9518-4037-b96b-a1490082fb04","Type":"ContainerStarted","Data":"8269b292deaebc44d46a82bcbd5a9052216ab55fce7ada8f29a4aafd48fa5ccb"} Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.740452 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58bbd77a-9518-4037-b96b-a1490082fb04","Type":"ContainerStarted","Data":"1450a27e6740c0748d4cf938867267b47c8cf60f863dd86e12df098c85f250cc"} Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.760732 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.760717717 podStartE2EDuration="2.760717717s" podCreationTimestamp="2026-01-24 00:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:25:05.757239711 +0000 UTC m=+1289.787210702" watchObservedRunningTime="2026-01-24 00:25:05.760717717 +0000 UTC m=+1289.790688718" Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.774636 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.928757 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-49926"] Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.929919 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.937985 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-49926"] Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.940076 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Jan 24 00:25:05 crc kubenswrapper[4676]: I0124 00:25:05.940471 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.041627 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.041675 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.078520 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.078598 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-scripts\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.078786 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-config-data\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.078951 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5hxl\" (UniqueName: \"kubernetes.io/projected/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-kube-api-access-q5hxl\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.180524 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5hxl\" (UniqueName: \"kubernetes.io/projected/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-kube-api-access-q5hxl\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.180678 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.180699 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-scripts\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.180731 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-config-data\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.186113 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-config-data\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.186690 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.190990 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-scripts\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.200277 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5hxl\" (UniqueName: \"kubernetes.io/projected/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-kube-api-access-q5hxl\") pod \"nova-cell1-cell-mapping-49926\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.253623 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.744367 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-49926"] Jan 24 00:25:06 crc kubenswrapper[4676]: W0124 00:25:06.756906 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e6d63aa_e36a_4ef7_b50d_255b44e72c20.slice/crio-11b0104944bb8c5aff79330efedf762b41dcc22f00267801cc45a32cf1468cf9 WatchSource:0}: Error finding container 11b0104944bb8c5aff79330efedf762b41dcc22f00267801cc45a32cf1468cf9: Status 404 returned error can't find the container with id 11b0104944bb8c5aff79330efedf762b41dcc22f00267801cc45a32cf1468cf9 Jan 24 00:25:06 crc kubenswrapper[4676]: I0124 00:25:06.758421 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58bbd77a-9518-4037-b96b-a1490082fb04","Type":"ContainerStarted","Data":"5a9e6225549dfb21345b08b2519c115a0add95915d84a2fb7b60cccef323a6c6"} Jan 24 00:25:07 crc kubenswrapper[4676]: I0124 00:25:07.498894 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:25:07 crc kubenswrapper[4676]: I0124 00:25:07.571552 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-72qp2"] Jan 24 00:25:07 crc kubenswrapper[4676]: I0124 00:25:07.572036 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" podUID="b8c07196-aa1d-4d14-bc9f-6aec4de13853" containerName="dnsmasq-dns" containerID="cri-o://2dc83da30d1c8b3f8b6fad2b698ea560b01dbdda3cf6667c49b816b31947008d" gracePeriod=10 Jan 24 00:25:07 crc kubenswrapper[4676]: I0124 00:25:07.782902 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-49926" event={"ID":"2e6d63aa-e36a-4ef7-b50d-255b44e72c20","Type":"ContainerStarted","Data":"d43689494700a5ceee020f09f4ca2f2fd150c60e2e3c7e0d6cd30145e29f275e"} Jan 24 00:25:07 crc kubenswrapper[4676]: I0124 00:25:07.782942 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-49926" event={"ID":"2e6d63aa-e36a-4ef7-b50d-255b44e72c20","Type":"ContainerStarted","Data":"11b0104944bb8c5aff79330efedf762b41dcc22f00267801cc45a32cf1468cf9"} Jan 24 00:25:07 crc kubenswrapper[4676]: I0124 00:25:07.793836 4676 generic.go:334] "Generic (PLEG): container finished" podID="b8c07196-aa1d-4d14-bc9f-6aec4de13853" containerID="2dc83da30d1c8b3f8b6fad2b698ea560b01dbdda3cf6667c49b816b31947008d" exitCode=0 Jan 24 00:25:07 crc kubenswrapper[4676]: I0124 00:25:07.793904 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" event={"ID":"b8c07196-aa1d-4d14-bc9f-6aec4de13853","Type":"ContainerDied","Data":"2dc83da30d1c8b3f8b6fad2b698ea560b01dbdda3cf6667c49b816b31947008d"} Jan 24 00:25:07 crc kubenswrapper[4676]: I0124 00:25:07.814709 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-49926" podStartSLOduration=2.814689528 podStartE2EDuration="2.814689528s" podCreationTimestamp="2026-01-24 00:25:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:25:07.811761858 +0000 UTC m=+1291.841732859" watchObservedRunningTime="2026-01-24 00:25:07.814689528 +0000 UTC m=+1291.844660529" Jan 24 00:25:07 crc kubenswrapper[4676]: I0124 00:25:07.816044 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58bbd77a-9518-4037-b96b-a1490082fb04","Type":"ContainerStarted","Data":"5a397a6753309135105283265cae4b39c50a6d461b49380c1661be0d1ddb8dee"} Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.161787 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.330115 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-svc\") pod \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.330177 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-config\") pod \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.330316 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2kxc\" (UniqueName: \"kubernetes.io/projected/b8c07196-aa1d-4d14-bc9f-6aec4de13853-kube-api-access-c2kxc\") pod \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.330333 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-sb\") pod \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.330356 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-swift-storage-0\") pod \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.330418 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-nb\") pod \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\" (UID: \"b8c07196-aa1d-4d14-bc9f-6aec4de13853\") " Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.355883 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8c07196-aa1d-4d14-bc9f-6aec4de13853-kube-api-access-c2kxc" (OuterVolumeSpecName: "kube-api-access-c2kxc") pod "b8c07196-aa1d-4d14-bc9f-6aec4de13853" (UID: "b8c07196-aa1d-4d14-bc9f-6aec4de13853"). InnerVolumeSpecName "kube-api-access-c2kxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.409021 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b8c07196-aa1d-4d14-bc9f-6aec4de13853" (UID: "b8c07196-aa1d-4d14-bc9f-6aec4de13853"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.410302 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b8c07196-aa1d-4d14-bc9f-6aec4de13853" (UID: "b8c07196-aa1d-4d14-bc9f-6aec4de13853"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.410798 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b8c07196-aa1d-4d14-bc9f-6aec4de13853" (UID: "b8c07196-aa1d-4d14-bc9f-6aec4de13853"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.413345 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-config" (OuterVolumeSpecName: "config") pod "b8c07196-aa1d-4d14-bc9f-6aec4de13853" (UID: "b8c07196-aa1d-4d14-bc9f-6aec4de13853"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.433424 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b8c07196-aa1d-4d14-bc9f-6aec4de13853" (UID: "b8c07196-aa1d-4d14-bc9f-6aec4de13853"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.434207 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2kxc\" (UniqueName: \"kubernetes.io/projected/b8c07196-aa1d-4d14-bc9f-6aec4de13853-kube-api-access-c2kxc\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.434240 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.434249 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.434257 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.434266 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.434274 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8c07196-aa1d-4d14-bc9f-6aec4de13853-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.827818 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" event={"ID":"b8c07196-aa1d-4d14-bc9f-6aec4de13853","Type":"ContainerDied","Data":"149be033f93bb329597c455a0e6c6a8a209721f648e144a6195c8b2dcfc6602d"} Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.827842 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-72qp2" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.827880 4676 scope.go:117] "RemoveContainer" containerID="2dc83da30d1c8b3f8b6fad2b698ea560b01dbdda3cf6667c49b816b31947008d" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.857426 4676 scope.go:117] "RemoveContainer" containerID="ab4cecb7436eb3bdc8075e2778a585be7c7bc22f8b1112d18ec061da1496853f" Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.892596 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-72qp2"] Jan 24 00:25:08 crc kubenswrapper[4676]: I0124 00:25:08.923299 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-72qp2"] Jan 24 00:25:09 crc kubenswrapper[4676]: I0124 00:25:09.839921 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"58bbd77a-9518-4037-b96b-a1490082fb04","Type":"ContainerStarted","Data":"795b1c1ddd3e5518853d874ea4d59834c66b9a6cf52194dfcd4ecebaabca0ce1"} Jan 24 00:25:09 crc kubenswrapper[4676]: I0124 00:25:09.840092 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 24 00:25:09 crc kubenswrapper[4676]: I0124 00:25:09.861321 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.118502461 podStartE2EDuration="6.861306444s" podCreationTimestamp="2026-01-24 00:25:03 +0000 UTC" firstStartedPulling="2026-01-24 00:25:04.822349627 +0000 UTC m=+1288.852320628" lastFinishedPulling="2026-01-24 00:25:08.56515361 +0000 UTC m=+1292.595124611" observedRunningTime="2026-01-24 00:25:09.859831539 +0000 UTC m=+1293.889802560" watchObservedRunningTime="2026-01-24 00:25:09.861306444 +0000 UTC m=+1293.891277445" Jan 24 00:25:10 crc kubenswrapper[4676]: I0124 00:25:10.264660 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8c07196-aa1d-4d14-bc9f-6aec4de13853" path="/var/lib/kubelet/pods/b8c07196-aa1d-4d14-bc9f-6aec4de13853/volumes" Jan 24 00:25:12 crc kubenswrapper[4676]: I0124 00:25:12.889370 4676 generic.go:334] "Generic (PLEG): container finished" podID="2e6d63aa-e36a-4ef7-b50d-255b44e72c20" containerID="d43689494700a5ceee020f09f4ca2f2fd150c60e2e3c7e0d6cd30145e29f275e" exitCode=0 Jan 24 00:25:12 crc kubenswrapper[4676]: I0124 00:25:12.889470 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-49926" event={"ID":"2e6d63aa-e36a-4ef7-b50d-255b44e72c20","Type":"ContainerDied","Data":"d43689494700a5ceee020f09f4ca2f2fd150c60e2e3c7e0d6cd30145e29f275e"} Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.127953 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.128304 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.332446 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.448600 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5hxl\" (UniqueName: \"kubernetes.io/projected/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-kube-api-access-q5hxl\") pod \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.449157 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-config-data\") pod \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.449270 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-combined-ca-bundle\") pod \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.449420 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-scripts\") pod \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\" (UID: \"2e6d63aa-e36a-4ef7-b50d-255b44e72c20\") " Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.468494 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-scripts" (OuterVolumeSpecName: "scripts") pod "2e6d63aa-e36a-4ef7-b50d-255b44e72c20" (UID: "2e6d63aa-e36a-4ef7-b50d-255b44e72c20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.470587 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-kube-api-access-q5hxl" (OuterVolumeSpecName: "kube-api-access-q5hxl") pod "2e6d63aa-e36a-4ef7-b50d-255b44e72c20" (UID: "2e6d63aa-e36a-4ef7-b50d-255b44e72c20"). InnerVolumeSpecName "kube-api-access-q5hxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.479621 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e6d63aa-e36a-4ef7-b50d-255b44e72c20" (UID: "2e6d63aa-e36a-4ef7-b50d-255b44e72c20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.482552 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-config-data" (OuterVolumeSpecName: "config-data") pod "2e6d63aa-e36a-4ef7-b50d-255b44e72c20" (UID: "2e6d63aa-e36a-4ef7-b50d-255b44e72c20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.551252 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.551280 4676 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-scripts\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.551290 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5hxl\" (UniqueName: \"kubernetes.io/projected/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-kube-api-access-q5hxl\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.551300 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6d63aa-e36a-4ef7-b50d-255b44e72c20-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.916962 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-49926" event={"ID":"2e6d63aa-e36a-4ef7-b50d-255b44e72c20","Type":"ContainerDied","Data":"11b0104944bb8c5aff79330efedf762b41dcc22f00267801cc45a32cf1468cf9"} Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.917006 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11b0104944bb8c5aff79330efedf762b41dcc22f00267801cc45a32cf1468cf9" Jan 24 00:25:14 crc kubenswrapper[4676]: I0124 00:25:14.917069 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-49926" Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.034012 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.035499 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.046631 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.109776 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.112219 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-api" containerID="cri-o://01fc81c9856cfd2ef13df58946b52afdb5cba674fdb4423a36d9a11f6c166109" gracePeriod=30 Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.112238 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-log" containerID="cri-o://71c7f3216e344862d8255e5daeef618a41a9f2fd2e7332ca45fe189abdb2cd5c" gracePeriod=30 Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.132428 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.132443 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.148766 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.149090 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="32d678d8-bc5d-4993-afc3-c6700d53b00e" containerName="nova-scheduler-scheduler" containerID="cri-o://9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" gracePeriod=30 Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.168127 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:25:15 crc kubenswrapper[4676]: E0124 00:25:15.649791 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 24 00:25:15 crc kubenswrapper[4676]: E0124 00:25:15.652273 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 24 00:25:15 crc kubenswrapper[4676]: E0124 00:25:15.653526 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 24 00:25:15 crc kubenswrapper[4676]: E0124 00:25:15.653598 4676 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="32d678d8-bc5d-4993-afc3-c6700d53b00e" containerName="nova-scheduler-scheduler" Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.928028 4676 generic.go:334] "Generic (PLEG): container finished" podID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerID="71c7f3216e344862d8255e5daeef618a41a9f2fd2e7332ca45fe189abdb2cd5c" exitCode=143 Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.928461 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace","Type":"ContainerDied","Data":"71c7f3216e344862d8255e5daeef618a41a9f2fd2e7332ca45fe189abdb2cd5c"} Jan 24 00:25:15 crc kubenswrapper[4676]: I0124 00:25:15.940720 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 24 00:25:16 crc kubenswrapper[4676]: I0124 00:25:16.936682 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-log" containerID="cri-o://dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1" gracePeriod=30 Jan 24 00:25:16 crc kubenswrapper[4676]: I0124 00:25:16.937114 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-metadata" containerID="cri-o://069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b" gracePeriod=30 Jan 24 00:25:17 crc kubenswrapper[4676]: I0124 00:25:17.950251 4676 generic.go:334] "Generic (PLEG): container finished" podID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerID="dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1" exitCode=143 Jan 24 00:25:17 crc kubenswrapper[4676]: I0124 00:25:17.950313 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6784012-a1a0-4647-929f-50788a8dc5bb","Type":"ContainerDied","Data":"dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1"} Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.092671 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:41200->10.217.0.195:8775: read: connection reset by peer" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.092746 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.195:8775/\": read tcp 10.217.0.2:41204->10.217.0.195:8775: read: connection reset by peer" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.554734 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:25:20 crc kubenswrapper[4676]: E0124 00:25:20.650002 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142 is running failed: container process not found" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 24 00:25:20 crc kubenswrapper[4676]: E0124 00:25:20.651231 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142 is running failed: container process not found" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 24 00:25:20 crc kubenswrapper[4676]: E0124 00:25:20.651474 4676 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142 is running failed: container process not found" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 24 00:25:20 crc kubenswrapper[4676]: E0124 00:25:20.651498 4676 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="32d678d8-bc5d-4993-afc3-c6700d53b00e" containerName="nova-scheduler-scheduler" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.660775 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.716211 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-combined-ca-bundle\") pod \"d6784012-a1a0-4647-929f-50788a8dc5bb\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.716338 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-config-data\") pod \"d6784012-a1a0-4647-929f-50788a8dc5bb\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.716365 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-nova-metadata-tls-certs\") pod \"d6784012-a1a0-4647-929f-50788a8dc5bb\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.717802 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4dfw\" (UniqueName: \"kubernetes.io/projected/d6784012-a1a0-4647-929f-50788a8dc5bb-kube-api-access-p4dfw\") pod \"d6784012-a1a0-4647-929f-50788a8dc5bb\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.717836 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6784012-a1a0-4647-929f-50788a8dc5bb-logs\") pod \"d6784012-a1a0-4647-929f-50788a8dc5bb\" (UID: \"d6784012-a1a0-4647-929f-50788a8dc5bb\") " Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.727870 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6784012-a1a0-4647-929f-50788a8dc5bb-logs" (OuterVolumeSpecName: "logs") pod "d6784012-a1a0-4647-929f-50788a8dc5bb" (UID: "d6784012-a1a0-4647-929f-50788a8dc5bb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.753183 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-config-data" (OuterVolumeSpecName: "config-data") pod "d6784012-a1a0-4647-929f-50788a8dc5bb" (UID: "d6784012-a1a0-4647-929f-50788a8dc5bb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.761223 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6784012-a1a0-4647-929f-50788a8dc5bb-kube-api-access-p4dfw" (OuterVolumeSpecName: "kube-api-access-p4dfw") pod "d6784012-a1a0-4647-929f-50788a8dc5bb" (UID: "d6784012-a1a0-4647-929f-50788a8dc5bb"). InnerVolumeSpecName "kube-api-access-p4dfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.804488 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6784012-a1a0-4647-929f-50788a8dc5bb" (UID: "d6784012-a1a0-4647-929f-50788a8dc5bb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.806530 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d6784012-a1a0-4647-929f-50788a8dc5bb" (UID: "d6784012-a1a0-4647-929f-50788a8dc5bb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.822516 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vr9g\" (UniqueName: \"kubernetes.io/projected/32d678d8-bc5d-4993-afc3-c6700d53b00e-kube-api-access-5vr9g\") pod \"32d678d8-bc5d-4993-afc3-c6700d53b00e\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.822576 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-combined-ca-bundle\") pod \"32d678d8-bc5d-4993-afc3-c6700d53b00e\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.822670 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-config-data\") pod \"32d678d8-bc5d-4993-afc3-c6700d53b00e\" (UID: \"32d678d8-bc5d-4993-afc3-c6700d53b00e\") " Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.823018 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.823030 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.823042 4676 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6784012-a1a0-4647-929f-50788a8dc5bb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.823052 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4dfw\" (UniqueName: \"kubernetes.io/projected/d6784012-a1a0-4647-929f-50788a8dc5bb-kube-api-access-p4dfw\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.823080 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6784012-a1a0-4647-929f-50788a8dc5bb-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.842154 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32d678d8-bc5d-4993-afc3-c6700d53b00e-kube-api-access-5vr9g" (OuterVolumeSpecName: "kube-api-access-5vr9g") pod "32d678d8-bc5d-4993-afc3-c6700d53b00e" (UID: "32d678d8-bc5d-4993-afc3-c6700d53b00e"). InnerVolumeSpecName "kube-api-access-5vr9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.862733 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32d678d8-bc5d-4993-afc3-c6700d53b00e" (UID: "32d678d8-bc5d-4993-afc3-c6700d53b00e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.862863 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-config-data" (OuterVolumeSpecName: "config-data") pod "32d678d8-bc5d-4993-afc3-c6700d53b00e" (UID: "32d678d8-bc5d-4993-afc3-c6700d53b00e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.924774 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vr9g\" (UniqueName: \"kubernetes.io/projected/32d678d8-bc5d-4993-afc3-c6700d53b00e-kube-api-access-5vr9g\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.924801 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.924810 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/32d678d8-bc5d-4993-afc3-c6700d53b00e-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.992221 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6784012-a1a0-4647-929f-50788a8dc5bb","Type":"ContainerDied","Data":"069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b"} Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.992251 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.992264 4676 scope.go:117] "RemoveContainer" containerID="069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b" Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.992169 4676 generic.go:334] "Generic (PLEG): container finished" podID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerID="069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b" exitCode=0 Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.992715 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d6784012-a1a0-4647-929f-50788a8dc5bb","Type":"ContainerDied","Data":"01fe7e14f549bff2abe35a4d5403466f7bbd36dfe57daa6ef4e421db5dcd7cf7"} Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.997367 4676 generic.go:334] "Generic (PLEG): container finished" podID="32d678d8-bc5d-4993-afc3-c6700d53b00e" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" exitCode=0 Jan 24 00:25:20 crc kubenswrapper[4676]: I0124 00:25:20.997471 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.001715 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"32d678d8-bc5d-4993-afc3-c6700d53b00e","Type":"ContainerDied","Data":"9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142"} Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.001768 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"32d678d8-bc5d-4993-afc3-c6700d53b00e","Type":"ContainerDied","Data":"d01bc4c049fe259a880c270ae5608bf35ba2c89a0a3367424a3555670349c800"} Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.006038 4676 generic.go:334] "Generic (PLEG): container finished" podID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerID="01fc81c9856cfd2ef13df58946b52afdb5cba674fdb4423a36d9a11f6c166109" exitCode=0 Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.006099 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace","Type":"ContainerDied","Data":"01fc81c9856cfd2ef13df58946b52afdb5cba674fdb4423a36d9a11f6c166109"} Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.044831 4676 scope.go:117] "RemoveContainer" containerID="dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.054394 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.068869 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.086669 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.091075 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100044 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.100370 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e6d63aa-e36a-4ef7-b50d-255b44e72c20" containerName="nova-manage" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100396 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e6d63aa-e36a-4ef7-b50d-255b44e72c20" containerName="nova-manage" Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.100410 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d678d8-bc5d-4993-afc3-c6700d53b00e" containerName="nova-scheduler-scheduler" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100416 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d678d8-bc5d-4993-afc3-c6700d53b00e" containerName="nova-scheduler-scheduler" Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.100441 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c07196-aa1d-4d14-bc9f-6aec4de13853" containerName="dnsmasq-dns" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100448 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c07196-aa1d-4d14-bc9f-6aec4de13853" containerName="dnsmasq-dns" Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.100458 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-log" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100463 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-log" Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.100478 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-metadata" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100483 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-metadata" Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.100491 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8c07196-aa1d-4d14-bc9f-6aec4de13853" containerName="init" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100497 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8c07196-aa1d-4d14-bc9f-6aec4de13853" containerName="init" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100663 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8c07196-aa1d-4d14-bc9f-6aec4de13853" containerName="dnsmasq-dns" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100673 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="32d678d8-bc5d-4993-afc3-c6700d53b00e" containerName="nova-scheduler-scheduler" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100683 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-log" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100692 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e6d63aa-e36a-4ef7-b50d-255b44e72c20" containerName="nova-manage" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.100701 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" containerName="nova-metadata-metadata" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.101624 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.102319 4676 scope.go:117] "RemoveContainer" containerID="069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b" Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.103095 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b\": container with ID starting with 069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b not found: ID does not exist" containerID="069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.103196 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b"} err="failed to get container status \"069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b\": rpc error: code = NotFound desc = could not find container \"069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b\": container with ID starting with 069d417103a5b3d1a63ca0a83a7bce5e55fb783cc6b5b9635b78d676f26ab99b not found: ID does not exist" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.103272 4676 scope.go:117] "RemoveContainer" containerID="dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1" Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.104273 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1\": container with ID starting with dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1 not found: ID does not exist" containerID="dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.104314 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1"} err="failed to get container status \"dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1\": rpc error: code = NotFound desc = could not find container \"dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1\": container with ID starting with dc7b56efe46aa7c7308ac62b0923d3c3d5f8466d0d2cfaa919b9e8096c2effb1 not found: ID does not exist" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.104346 4676 scope.go:117] "RemoveContainer" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.105215 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.107040 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.110014 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.117707 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.118326 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-log" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.118429 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-log" Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.118532 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-api" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.118596 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-api" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.118828 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-api" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.118913 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" containerName="nova-api-log" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.120258 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.125975 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.137071 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.163767 4676 scope.go:117] "RemoveContainer" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" Jan 24 00:25:21 crc kubenswrapper[4676]: E0124 00:25:21.164232 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142\": container with ID starting with 9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142 not found: ID does not exist" containerID="9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.164275 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142"} err="failed to get container status \"9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142\": rpc error: code = NotFound desc = could not find container \"9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142\": container with ID starting with 9ba31632fec3a842045888270699fb415b184651ae3d470415114c55c80c8142 not found: ID does not exist" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.185204 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.229251 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-internal-tls-certs\") pod \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.229426 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-config-data\") pod \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.229495 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-logs\") pod \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.229539 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdbxs\" (UniqueName: \"kubernetes.io/projected/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-kube-api-access-cdbxs\") pod \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.229616 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-public-tls-certs\") pod \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.229738 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-combined-ca-bundle\") pod \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\" (UID: \"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace\") " Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.230237 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1085d48-78f1-4437-a518-239ba90b7c0b-logs\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.230303 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1085d48-78f1-4437-a518-239ba90b7c0b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.230414 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-logs" (OuterVolumeSpecName: "logs") pod "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" (UID: "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.230620 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcmvm\" (UniqueName: \"kubernetes.io/projected/8ff6f87b-a571-43da-9fbc-9203f0001771-kube-api-access-wcmvm\") pod \"nova-scheduler-0\" (UID: \"8ff6f87b-a571-43da-9fbc-9203f0001771\") " pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.230710 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff6f87b-a571-43da-9fbc-9203f0001771-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8ff6f87b-a571-43da-9fbc-9203f0001771\") " pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.230780 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1085d48-78f1-4437-a518-239ba90b7c0b-config-data\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.230819 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1085d48-78f1-4437-a518-239ba90b7c0b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.230889 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mslmt\" (UniqueName: \"kubernetes.io/projected/f1085d48-78f1-4437-a518-239ba90b7c0b-kube-api-access-mslmt\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.231152 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ff6f87b-a571-43da-9fbc-9203f0001771-config-data\") pod \"nova-scheduler-0\" (UID: \"8ff6f87b-a571-43da-9fbc-9203f0001771\") " pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.231800 4676 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-logs\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.236784 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-kube-api-access-cdbxs" (OuterVolumeSpecName: "kube-api-access-cdbxs") pod "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" (UID: "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace"). InnerVolumeSpecName "kube-api-access-cdbxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.264968 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" (UID: "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.280547 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-config-data" (OuterVolumeSpecName: "config-data") pod "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" (UID: "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.287903 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" (UID: "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.290491 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" (UID: "8df0611d-44c4-4791-b1ee-ebfd5e9c4ace"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.333764 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ff6f87b-a571-43da-9fbc-9203f0001771-config-data\") pod \"nova-scheduler-0\" (UID: \"8ff6f87b-a571-43da-9fbc-9203f0001771\") " pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.333827 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1085d48-78f1-4437-a518-239ba90b7c0b-logs\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.333923 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1085d48-78f1-4437-a518-239ba90b7c0b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334138 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcmvm\" (UniqueName: \"kubernetes.io/projected/8ff6f87b-a571-43da-9fbc-9203f0001771-kube-api-access-wcmvm\") pod \"nova-scheduler-0\" (UID: \"8ff6f87b-a571-43da-9fbc-9203f0001771\") " pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334230 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff6f87b-a571-43da-9fbc-9203f0001771-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8ff6f87b-a571-43da-9fbc-9203f0001771\") " pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334293 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1085d48-78f1-4437-a518-239ba90b7c0b-config-data\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334327 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1085d48-78f1-4437-a518-239ba90b7c0b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334391 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1085d48-78f1-4437-a518-239ba90b7c0b-logs\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334420 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mslmt\" (UniqueName: \"kubernetes.io/projected/f1085d48-78f1-4437-a518-239ba90b7c0b-kube-api-access-mslmt\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334563 4676 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334580 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334593 4676 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334605 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.334618 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdbxs\" (UniqueName: \"kubernetes.io/projected/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace-kube-api-access-cdbxs\") on node \"crc\" DevicePath \"\"" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.338555 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ff6f87b-a571-43da-9fbc-9203f0001771-config-data\") pod \"nova-scheduler-0\" (UID: \"8ff6f87b-a571-43da-9fbc-9203f0001771\") " pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.340026 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1085d48-78f1-4437-a518-239ba90b7c0b-config-data\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.340345 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1085d48-78f1-4437-a518-239ba90b7c0b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.346516 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ff6f87b-a571-43da-9fbc-9203f0001771-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8ff6f87b-a571-43da-9fbc-9203f0001771\") " pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.351788 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1085d48-78f1-4437-a518-239ba90b7c0b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.353904 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcmvm\" (UniqueName: \"kubernetes.io/projected/8ff6f87b-a571-43da-9fbc-9203f0001771-kube-api-access-wcmvm\") pod \"nova-scheduler-0\" (UID: \"8ff6f87b-a571-43da-9fbc-9203f0001771\") " pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.357812 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mslmt\" (UniqueName: \"kubernetes.io/projected/f1085d48-78f1-4437-a518-239ba90b7c0b-kube-api-access-mslmt\") pod \"nova-metadata-0\" (UID: \"f1085d48-78f1-4437-a518-239ba90b7c0b\") " pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.437809 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.511600 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 24 00:25:21 crc kubenswrapper[4676]: W0124 00:25:21.944587 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1085d48_78f1_4437_a518_239ba90b7c0b.slice/crio-ca61a03b74a39bf03f1ded2a001e80bb2ece678117144da0b00b139baa25cdea WatchSource:0}: Error finding container ca61a03b74a39bf03f1ded2a001e80bb2ece678117144da0b00b139baa25cdea: Status 404 returned error can't find the container with id ca61a03b74a39bf03f1ded2a001e80bb2ece678117144da0b00b139baa25cdea Jan 24 00:25:21 crc kubenswrapper[4676]: I0124 00:25:21.950870 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.014019 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.040794 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8df0611d-44c4-4791-b1ee-ebfd5e9c4ace","Type":"ContainerDied","Data":"21b2d30f9d7e2c4a549d8c87dd63e0002ace80a317cb0d0205c94126fe98f6c1"} Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.040856 4676 scope.go:117] "RemoveContainer" containerID="01fc81c9856cfd2ef13df58946b52afdb5cba674fdb4423a36d9a11f6c166109" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.041068 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.043171 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1085d48-78f1-4437-a518-239ba90b7c0b","Type":"ContainerStarted","Data":"ca61a03b74a39bf03f1ded2a001e80bb2ece678117144da0b00b139baa25cdea"} Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.089539 4676 scope.go:117] "RemoveContainer" containerID="71c7f3216e344862d8255e5daeef618a41a9f2fd2e7332ca45fe189abdb2cd5c" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.109850 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.122271 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.135103 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.137056 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.140228 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.145487 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.146635 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.146964 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.249485 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.249549 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.249650 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff542eb3-0142-4d46-a6e6-2e89c73f5824-logs\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.249720 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-config-data\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.249767 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-public-tls-certs\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.249791 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfzr5\" (UniqueName: \"kubernetes.io/projected/ff542eb3-0142-4d46-a6e6-2e89c73f5824-kube-api-access-jfzr5\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.267812 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32d678d8-bc5d-4993-afc3-c6700d53b00e" path="/var/lib/kubelet/pods/32d678d8-bc5d-4993-afc3-c6700d53b00e/volumes" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.268495 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8df0611d-44c4-4791-b1ee-ebfd5e9c4ace" path="/var/lib/kubelet/pods/8df0611d-44c4-4791-b1ee-ebfd5e9c4ace/volumes" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.269204 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6784012-a1a0-4647-929f-50788a8dc5bb" path="/var/lib/kubelet/pods/d6784012-a1a0-4647-929f-50788a8dc5bb/volumes" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.351891 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff542eb3-0142-4d46-a6e6-2e89c73f5824-logs\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.351962 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-config-data\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.352010 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-public-tls-certs\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.352034 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfzr5\" (UniqueName: \"kubernetes.io/projected/ff542eb3-0142-4d46-a6e6-2e89c73f5824-kube-api-access-jfzr5\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.352459 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff542eb3-0142-4d46-a6e6-2e89c73f5824-logs\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.352664 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.353031 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.357349 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-internal-tls-certs\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.358598 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-config-data\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.359419 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.366722 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff542eb3-0142-4d46-a6e6-2e89c73f5824-public-tls-certs\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.372094 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfzr5\" (UniqueName: \"kubernetes.io/projected/ff542eb3-0142-4d46-a6e6-2e89c73f5824-kube-api-access-jfzr5\") pod \"nova-api-0\" (UID: \"ff542eb3-0142-4d46-a6e6-2e89c73f5824\") " pod="openstack/nova-api-0" Jan 24 00:25:22 crc kubenswrapper[4676]: I0124 00:25:22.627341 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 24 00:25:23 crc kubenswrapper[4676]: I0124 00:25:23.053113 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 24 00:25:23 crc kubenswrapper[4676]: I0124 00:25:23.064669 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1085d48-78f1-4437-a518-239ba90b7c0b","Type":"ContainerStarted","Data":"6bf98a0044126e70d8d0724db2f9a2c322af2b1541eaab3a04a5b65ffa344c6e"} Jan 24 00:25:23 crc kubenswrapper[4676]: I0124 00:25:23.064714 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1085d48-78f1-4437-a518-239ba90b7c0b","Type":"ContainerStarted","Data":"36247125dcfa3ca138a807b7c7c3bcb539eae76770c74e144f4f740836a033a2"} Jan 24 00:25:23 crc kubenswrapper[4676]: I0124 00:25:23.067798 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8ff6f87b-a571-43da-9fbc-9203f0001771","Type":"ContainerStarted","Data":"188dbba19d2bc9c27dd3a1c50427949a44481b8108236719f86cf0b3b1b7e55a"} Jan 24 00:25:23 crc kubenswrapper[4676]: I0124 00:25:23.067869 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8ff6f87b-a571-43da-9fbc-9203f0001771","Type":"ContainerStarted","Data":"9bc49e578ceca7e8e8b18c2ec8945fca3853f9ac77ae5798e51c50d5feb201dd"} Jan 24 00:25:23 crc kubenswrapper[4676]: I0124 00:25:23.092620 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.092602647 podStartE2EDuration="2.092602647s" podCreationTimestamp="2026-01-24 00:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:25:23.088737779 +0000 UTC m=+1307.118708820" watchObservedRunningTime="2026-01-24 00:25:23.092602647 +0000 UTC m=+1307.122573638" Jan 24 00:25:24 crc kubenswrapper[4676]: I0124 00:25:24.081357 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff542eb3-0142-4d46-a6e6-2e89c73f5824","Type":"ContainerStarted","Data":"98163d4d693f95146309125c2107c6529f89459431fd7e1fd9849d47d191bf29"} Jan 24 00:25:24 crc kubenswrapper[4676]: I0124 00:25:24.082860 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff542eb3-0142-4d46-a6e6-2e89c73f5824","Type":"ContainerStarted","Data":"640bb5433c9f92bd6425cafd54e570442e04f0073ec1546df3ee542a99930f01"} Jan 24 00:25:24 crc kubenswrapper[4676]: I0124 00:25:24.083005 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"ff542eb3-0142-4d46-a6e6-2e89c73f5824","Type":"ContainerStarted","Data":"bbe0d2c9e735026a39036967c182b36475355ec72fe3dd6eb21bf174887ad96f"} Jan 24 00:25:24 crc kubenswrapper[4676]: I0124 00:25:24.103033 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.103013582 podStartE2EDuration="3.103013582s" podCreationTimestamp="2026-01-24 00:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:25:23.114748812 +0000 UTC m=+1307.144719823" watchObservedRunningTime="2026-01-24 00:25:24.103013582 +0000 UTC m=+1308.132984573" Jan 24 00:25:24 crc kubenswrapper[4676]: I0124 00:25:24.107643 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.107634812 podStartE2EDuration="2.107634812s" podCreationTimestamp="2026-01-24 00:25:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:25:24.096645968 +0000 UTC m=+1308.126617009" watchObservedRunningTime="2026-01-24 00:25:24.107634812 +0000 UTC m=+1308.137605813" Jan 24 00:25:26 crc kubenswrapper[4676]: I0124 00:25:26.438127 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 24 00:25:26 crc kubenswrapper[4676]: I0124 00:25:26.438475 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 24 00:25:26 crc kubenswrapper[4676]: I0124 00:25:26.512424 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 24 00:25:31 crc kubenswrapper[4676]: I0124 00:25:31.438455 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 24 00:25:31 crc kubenswrapper[4676]: I0124 00:25:31.439437 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 24 00:25:31 crc kubenswrapper[4676]: I0124 00:25:31.512595 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 24 00:25:31 crc kubenswrapper[4676]: I0124 00:25:31.556658 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 24 00:25:32 crc kubenswrapper[4676]: I0124 00:25:32.208880 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 24 00:25:32 crc kubenswrapper[4676]: I0124 00:25:32.450521 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1085d48-78f1-4437-a518-239ba90b7c0b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:25:32 crc kubenswrapper[4676]: I0124 00:25:32.450525 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1085d48-78f1-4437-a518-239ba90b7c0b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.200:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:25:32 crc kubenswrapper[4676]: I0124 00:25:32.628978 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 24 00:25:32 crc kubenswrapper[4676]: I0124 00:25:32.629294 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 24 00:25:33 crc kubenswrapper[4676]: I0124 00:25:33.643531 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ff542eb3-0142-4d46-a6e6-2e89c73f5824" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:25:33 crc kubenswrapper[4676]: I0124 00:25:33.643599 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="ff542eb3-0142-4d46-a6e6-2e89c73f5824" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 24 00:25:34 crc kubenswrapper[4676]: I0124 00:25:34.226055 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 24 00:25:39 crc kubenswrapper[4676]: I0124 00:25:39.363749 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:25:39 crc kubenswrapper[4676]: I0124 00:25:39.364229 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:25:41 crc kubenswrapper[4676]: I0124 00:25:41.445832 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 24 00:25:41 crc kubenswrapper[4676]: I0124 00:25:41.447755 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 24 00:25:41 crc kubenswrapper[4676]: I0124 00:25:41.452469 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 24 00:25:42 crc kubenswrapper[4676]: I0124 00:25:42.277849 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 24 00:25:42 crc kubenswrapper[4676]: I0124 00:25:42.647134 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 24 00:25:42 crc kubenswrapper[4676]: I0124 00:25:42.647764 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 24 00:25:42 crc kubenswrapper[4676]: I0124 00:25:42.650937 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 24 00:25:42 crc kubenswrapper[4676]: I0124 00:25:42.668969 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 24 00:25:43 crc kubenswrapper[4676]: I0124 00:25:43.279431 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 24 00:25:43 crc kubenswrapper[4676]: I0124 00:25:43.290342 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 24 00:25:51 crc kubenswrapper[4676]: I0124 00:25:51.153106 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 24 00:25:51 crc kubenswrapper[4676]: I0124 00:25:51.994020 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 24 00:25:55 crc kubenswrapper[4676]: I0124 00:25:55.732973 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerName="rabbitmq" containerID="cri-o://7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385" gracePeriod=604796 Jan 24 00:25:55 crc kubenswrapper[4676]: I0124 00:25:55.802458 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Jan 24 00:25:56 crc kubenswrapper[4676]: I0124 00:25:56.611838 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="36558de2-6aac-43e9-832d-2f96c46e8152" containerName="rabbitmq" containerID="cri-o://028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9" gracePeriod=604796 Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.368369 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.439307 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/68d6466c-a6ff-40ba-952d-007b14efdfd3-erlang-cookie-secret\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.439739 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/68d6466c-a6ff-40ba-952d-007b14efdfd3-pod-info\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.439860 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-plugins\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.440019 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.440252 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-tls\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.440447 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-confd\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.440562 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.440763 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-config-data\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.440925 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-erlang-cookie\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.441076 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-server-conf\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.441225 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v959\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-kube-api-access-7v959\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.441325 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-plugins-conf\") pod \"68d6466c-a6ff-40ba-952d-007b14efdfd3\" (UID: \"68d6466c-a6ff-40ba-952d-007b14efdfd3\") " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.441968 4676 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.442585 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.442605 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.448470 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.448675 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/68d6466c-a6ff-40ba-952d-007b14efdfd3-pod-info" (OuterVolumeSpecName: "pod-info") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.481800 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-kube-api-access-7v959" (OuterVolumeSpecName: "kube-api-access-7v959") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "kube-api-access-7v959". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.482581 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.485590 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d6466c-a6ff-40ba-952d-007b14efdfd3-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.523906 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-config-data" (OuterVolumeSpecName: "config-data") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.540952 4676 generic.go:334] "Generic (PLEG): container finished" podID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerID="7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385" exitCode=0 Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.540991 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"68d6466c-a6ff-40ba-952d-007b14efdfd3","Type":"ContainerDied","Data":"7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385"} Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.541016 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"68d6466c-a6ff-40ba-952d-007b14efdfd3","Type":"ContainerDied","Data":"32030fde9c3a43091d81748f9987539fcf3f4b60bda9f5dc9e85b5f47719d7eb"} Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.541033 4676 scope.go:117] "RemoveContainer" containerID="7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.541165 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.546020 4676 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.546051 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.546063 4676 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.546074 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v959\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-kube-api-access-7v959\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.546086 4676 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.546095 4676 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/68d6466c-a6ff-40ba-952d-007b14efdfd3-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.546104 4676 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/68d6466c-a6ff-40ba-952d-007b14efdfd3-pod-info\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.546130 4676 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.582359 4676 scope.go:117] "RemoveContainer" containerID="0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.584864 4676 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.615539 4676 scope.go:117] "RemoveContainer" containerID="7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385" Jan 24 00:26:02 crc kubenswrapper[4676]: E0124 00:26:02.615959 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385\": container with ID starting with 7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385 not found: ID does not exist" containerID="7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.615986 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385"} err="failed to get container status \"7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385\": rpc error: code = NotFound desc = could not find container \"7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385\": container with ID starting with 7ab3a50e872eb2d62ccafb5072d9093f176eca7aee4a3935c8627a779d608385 not found: ID does not exist" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.616006 4676 scope.go:117] "RemoveContainer" containerID="0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723" Jan 24 00:26:02 crc kubenswrapper[4676]: E0124 00:26:02.616289 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723\": container with ID starting with 0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723 not found: ID does not exist" containerID="0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.616307 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723"} err="failed to get container status \"0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723\": rpc error: code = NotFound desc = could not find container \"0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723\": container with ID starting with 0db0998b0243f3673ede9e2b18bccbbf8c17216722eeabea36670a8396448723 not found: ID does not exist" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.625489 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-server-conf" (OuterVolumeSpecName: "server-conf") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.647470 4676 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.647501 4676 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/68d6466c-a6ff-40ba-952d-007b14efdfd3-server-conf\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.696783 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "68d6466c-a6ff-40ba-952d-007b14efdfd3" (UID: "68d6466c-a6ff-40ba-952d-007b14efdfd3"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.749008 4676 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/68d6466c-a6ff-40ba-952d-007b14efdfd3-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.887276 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.901838 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.913919 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 24 00:26:02 crc kubenswrapper[4676]: E0124 00:26:02.914680 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerName="rabbitmq" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.914697 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerName="rabbitmq" Jan 24 00:26:02 crc kubenswrapper[4676]: E0124 00:26:02.914717 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerName="setup-container" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.914723 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerName="setup-container" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.914902 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" containerName="rabbitmq" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.915785 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.917174 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.918048 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-nxmk6" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.918966 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.918968 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.919010 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.919043 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.919318 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 24 00:26:02 crc kubenswrapper[4676]: I0124 00:26:02.925752 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060260 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c162e478-58e3-4a83-97cb-29887613c1aa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060639 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c162e478-58e3-4a83-97cb-29887613c1aa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060675 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060716 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjv27\" (UniqueName: \"kubernetes.io/projected/c162e478-58e3-4a83-97cb-29887613c1aa-kube-api-access-vjv27\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060742 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060816 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c162e478-58e3-4a83-97cb-29887613c1aa-config-data\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060848 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060885 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c162e478-58e3-4a83-97cb-29887613c1aa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060905 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060922 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c162e478-58e3-4a83-97cb-29887613c1aa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.060949 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.162736 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c162e478-58e3-4a83-97cb-29887613c1aa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.163156 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.163274 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c162e478-58e3-4a83-97cb-29887613c1aa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.163436 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.163572 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c162e478-58e3-4a83-97cb-29887613c1aa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.163783 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c162e478-58e3-4a83-97cb-29887613c1aa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.163944 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.164104 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjv27\" (UniqueName: \"kubernetes.io/projected/c162e478-58e3-4a83-97cb-29887613c1aa-kube-api-access-vjv27\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.164246 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.164487 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c162e478-58e3-4a83-97cb-29887613c1aa-config-data\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.164629 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.164861 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.165955 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.166127 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.166155 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c162e478-58e3-4a83-97cb-29887613c1aa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.166666 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c162e478-58e3-4a83-97cb-29887613c1aa-config-data\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.166911 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c162e478-58e3-4a83-97cb-29887613c1aa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.171074 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.172091 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c162e478-58e3-4a83-97cb-29887613c1aa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.172773 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c162e478-58e3-4a83-97cb-29887613c1aa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.179165 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c162e478-58e3-4a83-97cb-29887613c1aa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.187691 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjv27\" (UniqueName: \"kubernetes.io/projected/c162e478-58e3-4a83-97cb-29887613c1aa-kube-api-access-vjv27\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.212479 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"c162e478-58e3-4a83-97cb-29887613c1aa\") " pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.233816 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.289030 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.471916 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-plugins-conf\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.471951 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.471994 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-plugins\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.472084 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36558de2-6aac-43e9-832d-2f96c46e8152-erlang-cookie-secret\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.472105 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdctt\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-kube-api-access-sdctt\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.472313 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-config-data\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.472808 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-confd\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.473242 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36558de2-6aac-43e9-832d-2f96c46e8152-pod-info\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.473260 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-server-conf\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.473369 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-erlang-cookie\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.473401 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-tls\") pod \"36558de2-6aac-43e9-832d-2f96c46e8152\" (UID: \"36558de2-6aac-43e9-832d-2f96c46e8152\") " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.473503 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.473881 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.474659 4676 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.474680 4676 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.474887 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.476883 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.483393 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/36558de2-6aac-43e9-832d-2f96c46e8152-pod-info" (OuterVolumeSpecName: "pod-info") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.483662 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36558de2-6aac-43e9-832d-2f96c46e8152-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.483714 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-kube-api-access-sdctt" (OuterVolumeSpecName: "kube-api-access-sdctt") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "kube-api-access-sdctt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.484525 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.510335 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-config-data" (OuterVolumeSpecName: "config-data") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.557090 4676 generic.go:334] "Generic (PLEG): container finished" podID="36558de2-6aac-43e9-832d-2f96c46e8152" containerID="028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9" exitCode=0 Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.557127 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"36558de2-6aac-43e9-832d-2f96c46e8152","Type":"ContainerDied","Data":"028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9"} Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.557150 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"36558de2-6aac-43e9-832d-2f96c46e8152","Type":"ContainerDied","Data":"a90a5dbbd754f4d91197a2d8faa756ac77a0f539415213117ebe9918b73147bc"} Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.557167 4676 scope.go:117] "RemoveContainer" containerID="028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.557280 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.564457 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-server-conf" (OuterVolumeSpecName: "server-conf") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.601310 4676 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/36558de2-6aac-43e9-832d-2f96c46e8152-pod-info\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.601538 4676 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-server-conf\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.601601 4676 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.601663 4676 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.601765 4676 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.601851 4676 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/36558de2-6aac-43e9-832d-2f96c46e8152-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.601916 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdctt\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-kube-api-access-sdctt\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.601994 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/36558de2-6aac-43e9-832d-2f96c46e8152-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.602603 4676 scope.go:117] "RemoveContainer" containerID="9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.624206 4676 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.631954 4676 scope.go:117] "RemoveContainer" containerID="028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9" Jan 24 00:26:03 crc kubenswrapper[4676]: E0124 00:26:03.632777 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9\": container with ID starting with 028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9 not found: ID does not exist" containerID="028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.632818 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9"} err="failed to get container status \"028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9\": rpc error: code = NotFound desc = could not find container \"028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9\": container with ID starting with 028693ff31c2a35e401cae916e148bd75fd1aa0c3135036bd6feedcc4c32fed9 not found: ID does not exist" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.632846 4676 scope.go:117] "RemoveContainer" containerID="9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960" Jan 24 00:26:03 crc kubenswrapper[4676]: E0124 00:26:03.633345 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960\": container with ID starting with 9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960 not found: ID does not exist" containerID="9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.633534 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960"} err="failed to get container status \"9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960\": rpc error: code = NotFound desc = could not find container \"9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960\": container with ID starting with 9d758f8c9e92f01e23b0692665a48971ec4eb0595df60b78749bc71450ba8960 not found: ID does not exist" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.656646 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "36558de2-6aac-43e9-832d-2f96c46e8152" (UID: "36558de2-6aac-43e9-832d-2f96c46e8152"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.703873 4676 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.703909 4676 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/36558de2-6aac-43e9-832d-2f96c46e8152-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.810031 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.908447 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.921513 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.939352 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 24 00:26:03 crc kubenswrapper[4676]: E0124 00:26:03.939766 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36558de2-6aac-43e9-832d-2f96c46e8152" containerName="setup-container" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.939782 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="36558de2-6aac-43e9-832d-2f96c46e8152" containerName="setup-container" Jan 24 00:26:03 crc kubenswrapper[4676]: E0124 00:26:03.939800 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36558de2-6aac-43e9-832d-2f96c46e8152" containerName="rabbitmq" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.939807 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="36558de2-6aac-43e9-832d-2f96c46e8152" containerName="rabbitmq" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.940026 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="36558de2-6aac-43e9-832d-2f96c46e8152" containerName="rabbitmq" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.941188 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.952703 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.952875 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.952973 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.953080 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-7x4sc" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.953183 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.953306 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.953484 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 24 00:26:03 crc kubenswrapper[4676]: I0124 00:26:03.986542 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.116900 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2df1d42-fa93-4771-ba77-1c27f820b298-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.116996 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2df1d42-fa93-4771-ba77-1c27f820b298-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.117036 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.117089 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.117159 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7d56\" (UniqueName: \"kubernetes.io/projected/a2df1d42-fa93-4771-ba77-1c27f820b298-kube-api-access-p7d56\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.117202 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.117269 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2df1d42-fa93-4771-ba77-1c27f820b298-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.117300 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2df1d42-fa93-4771-ba77-1c27f820b298-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.117330 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.117366 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.117428 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2df1d42-fa93-4771-ba77-1c27f820b298-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218494 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2df1d42-fa93-4771-ba77-1c27f820b298-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218543 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218576 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218614 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7d56\" (UniqueName: \"kubernetes.io/projected/a2df1d42-fa93-4771-ba77-1c27f820b298-kube-api-access-p7d56\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218642 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218679 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2df1d42-fa93-4771-ba77-1c27f820b298-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218698 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2df1d42-fa93-4771-ba77-1c27f820b298-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218717 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218743 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218770 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2df1d42-fa93-4771-ba77-1c27f820b298-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.218791 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2df1d42-fa93-4771-ba77-1c27f820b298-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.219887 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a2df1d42-fa93-4771-ba77-1c27f820b298-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.219979 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.220252 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.220361 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.220426 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a2df1d42-fa93-4771-ba77-1c27f820b298-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.221427 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a2df1d42-fa93-4771-ba77-1c27f820b298-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.224058 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a2df1d42-fa93-4771-ba77-1c27f820b298-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.226953 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.234830 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a2df1d42-fa93-4771-ba77-1c27f820b298-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.235442 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a2df1d42-fa93-4771-ba77-1c27f820b298-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.239894 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7d56\" (UniqueName: \"kubernetes.io/projected/a2df1d42-fa93-4771-ba77-1c27f820b298-kube-api-access-p7d56\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.267714 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36558de2-6aac-43e9-832d-2f96c46e8152" path="/var/lib/kubelet/pods/36558de2-6aac-43e9-832d-2f96c46e8152/volumes" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.269422 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68d6466c-a6ff-40ba-952d-007b14efdfd3" path="/var/lib/kubelet/pods/68d6466c-a6ff-40ba-952d-007b14efdfd3/volumes" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.271981 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a2df1d42-fa93-4771-ba77-1c27f820b298\") " pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.333116 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.569747 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c162e478-58e3-4a83-97cb-29887613c1aa","Type":"ContainerStarted","Data":"a86dae02bbfc444752a4a1ad59c4e336d8ba2bab252f75270770135809a7937e"} Jan 24 00:26:04 crc kubenswrapper[4676]: I0124 00:26:04.839195 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 24 00:26:04 crc kubenswrapper[4676]: W0124 00:26:04.840717 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2df1d42_fa93_4771_ba77_1c27f820b298.slice/crio-3e63c99875d3b0794d7835468269638558c3933816bb36b01c2e320b766fd256 WatchSource:0}: Error finding container 3e63c99875d3b0794d7835468269638558c3933816bb36b01c2e320b766fd256: Status 404 returned error can't find the container with id 3e63c99875d3b0794d7835468269638558c3933816bb36b01c2e320b766fd256 Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.020306 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-28gwc"] Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.022142 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.024508 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.050891 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-28gwc"] Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.163070 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.163129 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.163178 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-config\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.163203 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.163220 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbbrv\" (UniqueName: \"kubernetes.io/projected/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-kube-api-access-fbbrv\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.163288 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.163327 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-svc\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.265279 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.265367 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-svc\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.265474 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.265523 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.265585 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-config\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.265619 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.265643 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbbrv\" (UniqueName: \"kubernetes.io/projected/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-kube-api-access-fbbrv\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.266220 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.266250 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-svc\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.266792 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.266833 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.266871 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-config\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.267461 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.283591 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbbrv\" (UniqueName: \"kubernetes.io/projected/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-kube-api-access-fbbrv\") pod \"dnsmasq-dns-d558885bc-28gwc\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.350127 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.582336 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2df1d42-fa93-4771-ba77-1c27f820b298","Type":"ContainerStarted","Data":"3e63c99875d3b0794d7835468269638558c3933816bb36b01c2e320b766fd256"} Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.586450 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c162e478-58e3-4a83-97cb-29887613c1aa","Type":"ContainerStarted","Data":"e8dd570456f163b9ffc9b717def36eea3b68755520281c3ad977128488285d1c"} Jan 24 00:26:05 crc kubenswrapper[4676]: I0124 00:26:05.828349 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-28gwc"] Jan 24 00:26:06 crc kubenswrapper[4676]: I0124 00:26:06.595482 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2df1d42-fa93-4771-ba77-1c27f820b298","Type":"ContainerStarted","Data":"4622ad5ec55e893e88aecaf78be207183e4c4ff1e3645b53366aa4945663cc45"} Jan 24 00:26:06 crc kubenswrapper[4676]: I0124 00:26:06.597817 4676 generic.go:334] "Generic (PLEG): container finished" podID="cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" containerID="53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29" exitCode=0 Jan 24 00:26:06 crc kubenswrapper[4676]: I0124 00:26:06.597878 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-28gwc" event={"ID":"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa","Type":"ContainerDied","Data":"53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29"} Jan 24 00:26:06 crc kubenswrapper[4676]: I0124 00:26:06.598026 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-28gwc" event={"ID":"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa","Type":"ContainerStarted","Data":"47b70e12919d8c5f44b9279ac72d11f87aeeca3d747a40cf17fe0462e549f089"} Jan 24 00:26:07 crc kubenswrapper[4676]: I0124 00:26:07.607451 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-28gwc" event={"ID":"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa","Type":"ContainerStarted","Data":"2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3"} Jan 24 00:26:07 crc kubenswrapper[4676]: I0124 00:26:07.633270 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-28gwc" podStartSLOduration=3.633252618 podStartE2EDuration="3.633252618s" podCreationTimestamp="2026-01-24 00:26:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:26:07.626544164 +0000 UTC m=+1351.656515165" watchObservedRunningTime="2026-01-24 00:26:07.633252618 +0000 UTC m=+1351.663223619" Jan 24 00:26:08 crc kubenswrapper[4676]: I0124 00:26:08.620452 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:09 crc kubenswrapper[4676]: I0124 00:26:09.365222 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:26:09 crc kubenswrapper[4676]: I0124 00:26:09.365623 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.352918 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.435705 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-tkk74"] Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.435929 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" podUID="6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" containerName="dnsmasq-dns" containerID="cri-o://8189f6b4420a2e8c611c03a3298dcf14f84845a898137968cf5f45e6c0c37685" gracePeriod=10 Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.650466 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b865b64bc-drclt"] Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.652314 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.675315 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b865b64bc-drclt"] Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.694074 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.694113 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-config\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.694158 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-dns-svc\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.694200 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.694223 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqcd7\" (UniqueName: \"kubernetes.io/projected/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-kube-api-access-zqcd7\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.694262 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.694289 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.798183 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-dns-svc\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.798617 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.798669 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqcd7\" (UniqueName: \"kubernetes.io/projected/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-kube-api-access-zqcd7\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.798756 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.798812 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.798890 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.798921 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-config\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.799145 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-dns-svc\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.799473 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-dns-swift-storage-0\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.800019 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.800315 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-openstack-edpm-ipam\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.800993 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.804297 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-config\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.818710 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqcd7\" (UniqueName: \"kubernetes.io/projected/cdf394f7-5d67-4a0f-9644-82fe83a72e2d-kube-api-access-zqcd7\") pod \"dnsmasq-dns-6b865b64bc-drclt\" (UID: \"cdf394f7-5d67-4a0f-9644-82fe83a72e2d\") " pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.964757 4676 generic.go:334] "Generic (PLEG): container finished" podID="6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" containerID="8189f6b4420a2e8c611c03a3298dcf14f84845a898137968cf5f45e6c0c37685" exitCode=0 Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.964861 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" event={"ID":"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4","Type":"ContainerDied","Data":"8189f6b4420a2e8c611c03a3298dcf14f84845a898137968cf5f45e6c0c37685"} Jan 24 00:26:15 crc kubenswrapper[4676]: I0124 00:26:15.984084 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.081355 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.109958 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5hbp\" (UniqueName: \"kubernetes.io/projected/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-kube-api-access-s5hbp\") pod \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.110016 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-nb\") pod \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.110122 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-sb\") pod \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.110157 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-swift-storage-0\") pod \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.110204 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-config\") pod \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.110257 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-svc\") pod \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\" (UID: \"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4\") " Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.175637 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-kube-api-access-s5hbp" (OuterVolumeSpecName: "kube-api-access-s5hbp") pod "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" (UID: "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4"). InnerVolumeSpecName "kube-api-access-s5hbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.214886 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5hbp\" (UniqueName: \"kubernetes.io/projected/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-kube-api-access-s5hbp\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.345225 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" (UID: "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.346098 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.346274 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" (UID: "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.403698 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" (UID: "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.422147 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" (UID: "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.453614 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.453646 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.453657 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.482752 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-config" (OuterVolumeSpecName: "config") pod "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" (UID: "6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.506676 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b865b64bc-drclt"] Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.555308 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.977741 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" event={"ID":"6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4","Type":"ContainerDied","Data":"85858d9bee0a25807ae1aa70febaf41556800347d26ef61522b473ffb2662a4f"} Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.977767 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-tkk74" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.978278 4676 scope.go:117] "RemoveContainer" containerID="8189f6b4420a2e8c611c03a3298dcf14f84845a898137968cf5f45e6c0c37685" Jan 24 00:26:16 crc kubenswrapper[4676]: I0124 00:26:16.987128 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b865b64bc-drclt" event={"ID":"cdf394f7-5d67-4a0f-9644-82fe83a72e2d","Type":"ContainerStarted","Data":"570112f3c08284a28ce9d4293d6ec2bcf1e433dc8c0865373cc49543d157de29"} Jan 24 00:26:17 crc kubenswrapper[4676]: I0124 00:26:17.019202 4676 scope.go:117] "RemoveContainer" containerID="e9acf51095b7dd6a0ad30ae069a3cb2ce0ad324415a9a8bc556468f57ab1c34d" Jan 24 00:26:17 crc kubenswrapper[4676]: I0124 00:26:17.033859 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-tkk74"] Jan 24 00:26:17 crc kubenswrapper[4676]: I0124 00:26:17.048633 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-tkk74"] Jan 24 00:26:17 crc kubenswrapper[4676]: I0124 00:26:17.997093 4676 generic.go:334] "Generic (PLEG): container finished" podID="cdf394f7-5d67-4a0f-9644-82fe83a72e2d" containerID="195ec07999521e8d01c9bfdf5c88e4c63e302e89c9179fb6fbb32b00261a1c6d" exitCode=0 Jan 24 00:26:17 crc kubenswrapper[4676]: I0124 00:26:17.997188 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b865b64bc-drclt" event={"ID":"cdf394f7-5d67-4a0f-9644-82fe83a72e2d","Type":"ContainerDied","Data":"195ec07999521e8d01c9bfdf5c88e4c63e302e89c9179fb6fbb32b00261a1c6d"} Jan 24 00:26:18 crc kubenswrapper[4676]: I0124 00:26:18.267443 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" path="/var/lib/kubelet/pods/6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4/volumes" Jan 24 00:26:19 crc kubenswrapper[4676]: I0124 00:26:19.017008 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b865b64bc-drclt" event={"ID":"cdf394f7-5d67-4a0f-9644-82fe83a72e2d","Type":"ContainerStarted","Data":"09fcb71dd66ab87ad3c04aae1ffd1429983a5c4a5e0a4cf9884a9644c317d6ee"} Jan 24 00:26:19 crc kubenswrapper[4676]: I0124 00:26:19.018448 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:19 crc kubenswrapper[4676]: I0124 00:26:19.048387 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b865b64bc-drclt" podStartSLOduration=4.048359236 podStartE2EDuration="4.048359236s" podCreationTimestamp="2026-01-24 00:26:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:26:19.039383533 +0000 UTC m=+1363.069354544" watchObservedRunningTime="2026-01-24 00:26:19.048359236 +0000 UTC m=+1363.078330257" Jan 24 00:26:25 crc kubenswrapper[4676]: I0124 00:26:25.985709 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b865b64bc-drclt" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.073776 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-28gwc"] Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.079478 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-28gwc" podUID="cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" containerName="dnsmasq-dns" containerID="cri-o://2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3" gracePeriod=10 Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.531630 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.665528 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-openstack-edpm-ipam\") pod \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.665581 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-config\") pod \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.665625 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-nb\") pod \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.665661 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-svc\") pod \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.665692 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-sb\") pod \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.665726 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbbrv\" (UniqueName: \"kubernetes.io/projected/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-kube-api-access-fbbrv\") pod \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.665742 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-swift-storage-0\") pod \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\" (UID: \"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa\") " Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.679580 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-kube-api-access-fbbrv" (OuterVolumeSpecName: "kube-api-access-fbbrv") pod "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" (UID: "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa"). InnerVolumeSpecName "kube-api-access-fbbrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.724372 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" (UID: "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.730934 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-config" (OuterVolumeSpecName: "config") pod "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" (UID: "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.731713 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" (UID: "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.737775 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" (UID: "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.743626 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" (UID: "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.748075 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" (UID: "cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.767662 4676 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.767693 4676 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.767702 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.767712 4676 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.767720 4676 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.767729 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbbrv\" (UniqueName: \"kubernetes.io/projected/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-kube-api-access-fbbrv\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:26 crc kubenswrapper[4676]: I0124 00:26:26.767738 4676 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.099959 4676 generic.go:334] "Generic (PLEG): container finished" podID="cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" containerID="2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3" exitCode=0 Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.100018 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-28gwc" event={"ID":"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa","Type":"ContainerDied","Data":"2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3"} Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.100046 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-28gwc" Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.100087 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-28gwc" event={"ID":"cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa","Type":"ContainerDied","Data":"47b70e12919d8c5f44b9279ac72d11f87aeeca3d747a40cf17fe0462e549f089"} Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.100118 4676 scope.go:117] "RemoveContainer" containerID="2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3" Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.136339 4676 scope.go:117] "RemoveContainer" containerID="53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29" Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.156654 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-28gwc"] Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.162681 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-28gwc"] Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.174789 4676 scope.go:117] "RemoveContainer" containerID="2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3" Jan 24 00:26:27 crc kubenswrapper[4676]: E0124 00:26:27.175274 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3\": container with ID starting with 2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3 not found: ID does not exist" containerID="2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3" Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.175309 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3"} err="failed to get container status \"2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3\": rpc error: code = NotFound desc = could not find container \"2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3\": container with ID starting with 2c37d9e66e5004b7c1d2e801bd1fd56b36ccdd85608287fcda0faf0b75dfccc3 not found: ID does not exist" Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.175330 4676 scope.go:117] "RemoveContainer" containerID="53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29" Jan 24 00:26:27 crc kubenswrapper[4676]: E0124 00:26:27.175631 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29\": container with ID starting with 53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29 not found: ID does not exist" containerID="53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29" Jan 24 00:26:27 crc kubenswrapper[4676]: I0124 00:26:27.175651 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29"} err="failed to get container status \"53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29\": rpc error: code = NotFound desc = could not find container \"53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29\": container with ID starting with 53310433eaa7d3cfca542e7eeab8e064ca40864427d91517b1468dac71afcd29 not found: ID does not exist" Jan 24 00:26:28 crc kubenswrapper[4676]: I0124 00:26:28.271183 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" path="/var/lib/kubelet/pods/cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa/volumes" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.127991 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mfnp9"] Jan 24 00:26:35 crc kubenswrapper[4676]: E0124 00:26:35.130020 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" containerName="init" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.130136 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" containerName="init" Jan 24 00:26:35 crc kubenswrapper[4676]: E0124 00:26:35.130355 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" containerName="dnsmasq-dns" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.130480 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" containerName="dnsmasq-dns" Jan 24 00:26:35 crc kubenswrapper[4676]: E0124 00:26:35.130572 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" containerName="dnsmasq-dns" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.130664 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" containerName="dnsmasq-dns" Jan 24 00:26:35 crc kubenswrapper[4676]: E0124 00:26:35.130754 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" containerName="init" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.130834 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" containerName="init" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.131194 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a3231ad-bc2c-4a51-813d-cfd1a11c4fc4" containerName="dnsmasq-dns" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.131372 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd7a87a4-c4de-4fa5-ad41-c9f43405b2aa" containerName="dnsmasq-dns" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.133150 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.136863 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mfnp9"] Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.146876 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-utilities\") pod \"redhat-operators-mfnp9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.147092 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwdg8\" (UniqueName: \"kubernetes.io/projected/fde71182-237b-4e5e-9acd-7c063187e9d9-kube-api-access-cwdg8\") pod \"redhat-operators-mfnp9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.147303 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-catalog-content\") pod \"redhat-operators-mfnp9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.249775 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-catalog-content\") pod \"redhat-operators-mfnp9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.249899 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-utilities\") pod \"redhat-operators-mfnp9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.249932 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwdg8\" (UniqueName: \"kubernetes.io/projected/fde71182-237b-4e5e-9acd-7c063187e9d9-kube-api-access-cwdg8\") pod \"redhat-operators-mfnp9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.250583 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-catalog-content\") pod \"redhat-operators-mfnp9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.250607 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-utilities\") pod \"redhat-operators-mfnp9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.273154 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwdg8\" (UniqueName: \"kubernetes.io/projected/fde71182-237b-4e5e-9acd-7c063187e9d9-kube-api-access-cwdg8\") pod \"redhat-operators-mfnp9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.458046 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:35 crc kubenswrapper[4676]: I0124 00:26:35.905752 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mfnp9"] Jan 24 00:26:36 crc kubenswrapper[4676]: I0124 00:26:36.205018 4676 generic.go:334] "Generic (PLEG): container finished" podID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerID="b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5" exitCode=0 Jan 24 00:26:36 crc kubenswrapper[4676]: I0124 00:26:36.205185 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfnp9" event={"ID":"fde71182-237b-4e5e-9acd-7c063187e9d9","Type":"ContainerDied","Data":"b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5"} Jan 24 00:26:36 crc kubenswrapper[4676]: I0124 00:26:36.206456 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfnp9" event={"ID":"fde71182-237b-4e5e-9acd-7c063187e9d9","Type":"ContainerStarted","Data":"3a549d24e368d493375a69f5349ff827828a9aa43aab31f6ce9fe3d410a59e73"} Jan 24 00:26:38 crc kubenswrapper[4676]: I0124 00:26:38.225316 4676 generic.go:334] "Generic (PLEG): container finished" podID="c162e478-58e3-4a83-97cb-29887613c1aa" containerID="e8dd570456f163b9ffc9b717def36eea3b68755520281c3ad977128488285d1c" exitCode=0 Jan 24 00:26:38 crc kubenswrapper[4676]: I0124 00:26:38.225454 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c162e478-58e3-4a83-97cb-29887613c1aa","Type":"ContainerDied","Data":"e8dd570456f163b9ffc9b717def36eea3b68755520281c3ad977128488285d1c"} Jan 24 00:26:38 crc kubenswrapper[4676]: I0124 00:26:38.229211 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfnp9" event={"ID":"fde71182-237b-4e5e-9acd-7c063187e9d9","Type":"ContainerStarted","Data":"840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab"} Jan 24 00:26:38 crc kubenswrapper[4676]: I0124 00:26:38.233647 4676 generic.go:334] "Generic (PLEG): container finished" podID="a2df1d42-fa93-4771-ba77-1c27f820b298" containerID="4622ad5ec55e893e88aecaf78be207183e4c4ff1e3645b53366aa4945663cc45" exitCode=0 Jan 24 00:26:38 crc kubenswrapper[4676]: I0124 00:26:38.233703 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2df1d42-fa93-4771-ba77-1c27f820b298","Type":"ContainerDied","Data":"4622ad5ec55e893e88aecaf78be207183e4c4ff1e3645b53366aa4945663cc45"} Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.252791 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a2df1d42-fa93-4771-ba77-1c27f820b298","Type":"ContainerStarted","Data":"14f5b6f59284fc6ae60eca31583a70f5c98586f73492946534537fd1e26f3a40"} Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.253257 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.254901 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c162e478-58e3-4a83-97cb-29887613c1aa","Type":"ContainerStarted","Data":"16b670cb5e6da50b2aaba95552fa1fbf22bc6d2fdaafcc881aec602e2df9dbda"} Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.255263 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.288812 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.288791512 podStartE2EDuration="36.288791512s" podCreationTimestamp="2026-01-24 00:26:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:26:39.273242508 +0000 UTC m=+1383.303213509" watchObservedRunningTime="2026-01-24 00:26:39.288791512 +0000 UTC m=+1383.318762523" Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.305836 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.305818207 podStartE2EDuration="37.305818207s" podCreationTimestamp="2026-01-24 00:26:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 00:26:39.29753582 +0000 UTC m=+1383.327506821" watchObservedRunningTime="2026-01-24 00:26:39.305818207 +0000 UTC m=+1383.335789208" Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.363760 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.363819 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.363871 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.364744 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c0fdc4fa29c1a85e06a8e0b7d6899a3299af1dbe5c0f08e67bee66057d6c55e"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:26:39 crc kubenswrapper[4676]: I0124 00:26:39.364808 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://6c0fdc4fa29c1a85e06a8e0b7d6899a3299af1dbe5c0f08e67bee66057d6c55e" gracePeriod=600 Jan 24 00:26:41 crc kubenswrapper[4676]: I0124 00:26:41.273804 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="6c0fdc4fa29c1a85e06a8e0b7d6899a3299af1dbe5c0f08e67bee66057d6c55e" exitCode=0 Jan 24 00:26:41 crc kubenswrapper[4676]: I0124 00:26:41.273894 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"6c0fdc4fa29c1a85e06a8e0b7d6899a3299af1dbe5c0f08e67bee66057d6c55e"} Jan 24 00:26:41 crc kubenswrapper[4676]: I0124 00:26:41.274144 4676 scope.go:117] "RemoveContainer" containerID="f13744ab61f6ff84c30249dfd3e19836649d7bb6c4e4a3db144939c565fd684d" Jan 24 00:26:42 crc kubenswrapper[4676]: I0124 00:26:42.287840 4676 generic.go:334] "Generic (PLEG): container finished" podID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerID="840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab" exitCode=0 Jan 24 00:26:42 crc kubenswrapper[4676]: I0124 00:26:42.287894 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfnp9" event={"ID":"fde71182-237b-4e5e-9acd-7c063187e9d9","Type":"ContainerDied","Data":"840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab"} Jan 24 00:26:43 crc kubenswrapper[4676]: I0124 00:26:43.316160 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfnp9" event={"ID":"fde71182-237b-4e5e-9acd-7c063187e9d9","Type":"ContainerStarted","Data":"e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84"} Jan 24 00:26:43 crc kubenswrapper[4676]: I0124 00:26:43.325607 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751"} Jan 24 00:26:43 crc kubenswrapper[4676]: I0124 00:26:43.340676 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mfnp9" podStartSLOduration=1.671082754 podStartE2EDuration="8.34066231s" podCreationTimestamp="2026-01-24 00:26:35 +0000 UTC" firstStartedPulling="2026-01-24 00:26:36.206365253 +0000 UTC m=+1380.236336254" lastFinishedPulling="2026-01-24 00:26:42.875944799 +0000 UTC m=+1386.905915810" observedRunningTime="2026-01-24 00:26:43.336047359 +0000 UTC m=+1387.366018360" watchObservedRunningTime="2026-01-24 00:26:43.34066231 +0000 UTC m=+1387.370633311" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.474535 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd"] Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.476478 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.479992 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.480941 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.484668 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.485694 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd"] Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.485943 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.534587 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.534676 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.534705 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcrdb\" (UniqueName: \"kubernetes.io/projected/15cfbdc6-1f3b-49a5-8822-c4af1e686731-kube-api-access-vcrdb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.534730 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.637638 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.637725 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.637756 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcrdb\" (UniqueName: \"kubernetes.io/projected/15cfbdc6-1f3b-49a5-8822-c4af1e686731-kube-api-access-vcrdb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.637790 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.644404 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.644510 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.658067 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.658272 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcrdb\" (UniqueName: \"kubernetes.io/projected/15cfbdc6-1f3b-49a5-8822-c4af1e686731-kube-api-access-vcrdb\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:44 crc kubenswrapper[4676]: I0124 00:26:44.792407 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:26:45 crc kubenswrapper[4676]: I0124 00:26:45.437554 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd"] Jan 24 00:26:45 crc kubenswrapper[4676]: I0124 00:26:45.459564 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:45 crc kubenswrapper[4676]: I0124 00:26:45.459618 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:26:45 crc kubenswrapper[4676]: W0124 00:26:45.501950 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15cfbdc6_1f3b_49a5_8822_c4af1e686731.slice/crio-bad7f1fcff1a502c1a012676a123e6a6f137a9281f0ba2ea26bfe6c7df095d06 WatchSource:0}: Error finding container bad7f1fcff1a502c1a012676a123e6a6f137a9281f0ba2ea26bfe6c7df095d06: Status 404 returned error can't find the container with id bad7f1fcff1a502c1a012676a123e6a6f137a9281f0ba2ea26bfe6c7df095d06 Jan 24 00:26:46 crc kubenswrapper[4676]: I0124 00:26:46.360230 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" event={"ID":"15cfbdc6-1f3b-49a5-8822-c4af1e686731","Type":"ContainerStarted","Data":"bad7f1fcff1a502c1a012676a123e6a6f137a9281f0ba2ea26bfe6c7df095d06"} Jan 24 00:26:46 crc kubenswrapper[4676]: I0124 00:26:46.547116 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mfnp9" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="registry-server" probeResult="failure" output=< Jan 24 00:26:46 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:26:46 crc kubenswrapper[4676]: > Jan 24 00:26:53 crc kubenswrapper[4676]: I0124 00:26:53.244773 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 24 00:26:54 crc kubenswrapper[4676]: I0124 00:26:54.341923 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 24 00:26:56 crc kubenswrapper[4676]: I0124 00:26:56.537811 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mfnp9" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="registry-server" probeResult="failure" output=< Jan 24 00:26:56 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:26:56 crc kubenswrapper[4676]: > Jan 24 00:26:57 crc kubenswrapper[4676]: I0124 00:26:57.494975 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" event={"ID":"15cfbdc6-1f3b-49a5-8822-c4af1e686731","Type":"ContainerStarted","Data":"d3abe699a1e3161e3d65d105949518b56348bef4880f815312f858cc86bdc39a"} Jan 24 00:26:57 crc kubenswrapper[4676]: I0124 00:26:57.523992 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" podStartSLOduration=2.053822696 podStartE2EDuration="13.523973088s" podCreationTimestamp="2026-01-24 00:26:44 +0000 UTC" firstStartedPulling="2026-01-24 00:26:45.508075258 +0000 UTC m=+1389.538046259" lastFinishedPulling="2026-01-24 00:26:56.97822565 +0000 UTC m=+1401.008196651" observedRunningTime="2026-01-24 00:26:57.52191933 +0000 UTC m=+1401.551890351" watchObservedRunningTime="2026-01-24 00:26:57.523973088 +0000 UTC m=+1401.553944099" Jan 24 00:27:05 crc kubenswrapper[4676]: I0124 00:27:05.509993 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:27:05 crc kubenswrapper[4676]: I0124 00:27:05.572855 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:27:06 crc kubenswrapper[4676]: I0124 00:27:06.338170 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mfnp9"] Jan 24 00:27:06 crc kubenswrapper[4676]: I0124 00:27:06.577178 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mfnp9" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="registry-server" containerID="cri-o://e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84" gracePeriod=2 Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.085559 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.093035 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwdg8\" (UniqueName: \"kubernetes.io/projected/fde71182-237b-4e5e-9acd-7c063187e9d9-kube-api-access-cwdg8\") pod \"fde71182-237b-4e5e-9acd-7c063187e9d9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.093124 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-utilities\") pod \"fde71182-237b-4e5e-9acd-7c063187e9d9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.093154 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-catalog-content\") pod \"fde71182-237b-4e5e-9acd-7c063187e9d9\" (UID: \"fde71182-237b-4e5e-9acd-7c063187e9d9\") " Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.095048 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-utilities" (OuterVolumeSpecName: "utilities") pod "fde71182-237b-4e5e-9acd-7c063187e9d9" (UID: "fde71182-237b-4e5e-9acd-7c063187e9d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.105946 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fde71182-237b-4e5e-9acd-7c063187e9d9-kube-api-access-cwdg8" (OuterVolumeSpecName: "kube-api-access-cwdg8") pod "fde71182-237b-4e5e-9acd-7c063187e9d9" (UID: "fde71182-237b-4e5e-9acd-7c063187e9d9"). InnerVolumeSpecName "kube-api-access-cwdg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.199314 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwdg8\" (UniqueName: \"kubernetes.io/projected/fde71182-237b-4e5e-9acd-7c063187e9d9-kube-api-access-cwdg8\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.199362 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.252767 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fde71182-237b-4e5e-9acd-7c063187e9d9" (UID: "fde71182-237b-4e5e-9acd-7c063187e9d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.301719 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fde71182-237b-4e5e-9acd-7c063187e9d9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.589364 4676 generic.go:334] "Generic (PLEG): container finished" podID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerID="e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84" exitCode=0 Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.589404 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfnp9" event={"ID":"fde71182-237b-4e5e-9acd-7c063187e9d9","Type":"ContainerDied","Data":"e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84"} Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.589444 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mfnp9" event={"ID":"fde71182-237b-4e5e-9acd-7c063187e9d9","Type":"ContainerDied","Data":"3a549d24e368d493375a69f5349ff827828a9aa43aab31f6ce9fe3d410a59e73"} Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.589469 4676 scope.go:117] "RemoveContainer" containerID="e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.589465 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mfnp9" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.627663 4676 scope.go:117] "RemoveContainer" containerID="840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.630599 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mfnp9"] Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.639322 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mfnp9"] Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.650893 4676 scope.go:117] "RemoveContainer" containerID="b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.700415 4676 scope.go:117] "RemoveContainer" containerID="e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84" Jan 24 00:27:07 crc kubenswrapper[4676]: E0124 00:27:07.701000 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84\": container with ID starting with e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84 not found: ID does not exist" containerID="e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.701133 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84"} err="failed to get container status \"e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84\": rpc error: code = NotFound desc = could not find container \"e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84\": container with ID starting with e055fbf4304941e63aaafead3ca8b744c531ee533151d9a0bf016a4015b5bf84 not found: ID does not exist" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.701220 4676 scope.go:117] "RemoveContainer" containerID="840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab" Jan 24 00:27:07 crc kubenswrapper[4676]: E0124 00:27:07.701593 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab\": container with ID starting with 840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab not found: ID does not exist" containerID="840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.701624 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab"} err="failed to get container status \"840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab\": rpc error: code = NotFound desc = could not find container \"840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab\": container with ID starting with 840682d1bbe8575a5b2335cc56507fb17db2824f08a0d8e5489d08c656cbb9ab not found: ID does not exist" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.701645 4676 scope.go:117] "RemoveContainer" containerID="b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5" Jan 24 00:27:07 crc kubenswrapper[4676]: E0124 00:27:07.702669 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5\": container with ID starting with b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5 not found: ID does not exist" containerID="b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5" Jan 24 00:27:07 crc kubenswrapper[4676]: I0124 00:27:07.702754 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5"} err="failed to get container status \"b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5\": rpc error: code = NotFound desc = could not find container \"b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5\": container with ID starting with b4efdc97c998549b295a5310ebae1d71becdddc1045908ee456a712d7b3212b5 not found: ID does not exist" Jan 24 00:27:08 crc kubenswrapper[4676]: I0124 00:27:08.275069 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" path="/var/lib/kubelet/pods/fde71182-237b-4e5e-9acd-7c063187e9d9/volumes" Jan 24 00:27:09 crc kubenswrapper[4676]: I0124 00:27:09.614153 4676 generic.go:334] "Generic (PLEG): container finished" podID="15cfbdc6-1f3b-49a5-8822-c4af1e686731" containerID="d3abe699a1e3161e3d65d105949518b56348bef4880f815312f858cc86bdc39a" exitCode=0 Jan 24 00:27:09 crc kubenswrapper[4676]: I0124 00:27:09.614254 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" event={"ID":"15cfbdc6-1f3b-49a5-8822-c4af1e686731","Type":"ContainerDied","Data":"d3abe699a1e3161e3d65d105949518b56348bef4880f815312f858cc86bdc39a"} Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.075598 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.278017 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-ssh-key-openstack-edpm-ipam\") pod \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.278128 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcrdb\" (UniqueName: \"kubernetes.io/projected/15cfbdc6-1f3b-49a5-8822-c4af1e686731-kube-api-access-vcrdb\") pod \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.278196 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-repo-setup-combined-ca-bundle\") pod \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.279264 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-inventory\") pod \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\" (UID: \"15cfbdc6-1f3b-49a5-8822-c4af1e686731\") " Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.286316 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15cfbdc6-1f3b-49a5-8822-c4af1e686731-kube-api-access-vcrdb" (OuterVolumeSpecName: "kube-api-access-vcrdb") pod "15cfbdc6-1f3b-49a5-8822-c4af1e686731" (UID: "15cfbdc6-1f3b-49a5-8822-c4af1e686731"). InnerVolumeSpecName "kube-api-access-vcrdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.287501 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "15cfbdc6-1f3b-49a5-8822-c4af1e686731" (UID: "15cfbdc6-1f3b-49a5-8822-c4af1e686731"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.313594 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-inventory" (OuterVolumeSpecName: "inventory") pod "15cfbdc6-1f3b-49a5-8822-c4af1e686731" (UID: "15cfbdc6-1f3b-49a5-8822-c4af1e686731"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.315114 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "15cfbdc6-1f3b-49a5-8822-c4af1e686731" (UID: "15cfbdc6-1f3b-49a5-8822-c4af1e686731"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.382481 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.382519 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcrdb\" (UniqueName: \"kubernetes.io/projected/15cfbdc6-1f3b-49a5-8822-c4af1e686731-kube-api-access-vcrdb\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.382531 4676 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.382544 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15cfbdc6-1f3b-49a5-8822-c4af1e686731-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.637797 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" event={"ID":"15cfbdc6-1f3b-49a5-8822-c4af1e686731","Type":"ContainerDied","Data":"bad7f1fcff1a502c1a012676a123e6a6f137a9281f0ba2ea26bfe6c7df095d06"} Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.637843 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bad7f1fcff1a502c1a012676a123e6a6f137a9281f0ba2ea26bfe6c7df095d06" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.637925 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.781533 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8"] Jan 24 00:27:11 crc kubenswrapper[4676]: E0124 00:27:11.782673 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="extract-content" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.782722 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="extract-content" Jan 24 00:27:11 crc kubenswrapper[4676]: E0124 00:27:11.782769 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="extract-utilities" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.782784 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="extract-utilities" Jan 24 00:27:11 crc kubenswrapper[4676]: E0124 00:27:11.782815 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="registry-server" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.782828 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="registry-server" Jan 24 00:27:11 crc kubenswrapper[4676]: E0124 00:27:11.782876 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15cfbdc6-1f3b-49a5-8822-c4af1e686731" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.782894 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="15cfbdc6-1f3b-49a5-8822-c4af1e686731" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.783332 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="15cfbdc6-1f3b-49a5-8822-c4af1e686731" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.783372 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="fde71182-237b-4e5e-9acd-7c063187e9d9" containerName="registry-server" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.784534 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.789630 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qx6r8\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.789737 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slkrn\" (UniqueName: \"kubernetes.io/projected/19b712ec-28a7-419f-9f09-7d0b0ecbf747-kube-api-access-slkrn\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qx6r8\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.789938 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qx6r8\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.794069 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.794433 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.794500 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.795067 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.799612 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8"] Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.892141 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qx6r8\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.892281 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slkrn\" (UniqueName: \"kubernetes.io/projected/19b712ec-28a7-419f-9f09-7d0b0ecbf747-kube-api-access-slkrn\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qx6r8\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.892368 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qx6r8\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.898052 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qx6r8\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.899192 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qx6r8\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:11 crc kubenswrapper[4676]: I0124 00:27:11.914707 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slkrn\" (UniqueName: \"kubernetes.io/projected/19b712ec-28a7-419f-9f09-7d0b0ecbf747-kube-api-access-slkrn\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qx6r8\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:12 crc kubenswrapper[4676]: I0124 00:27:12.112016 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:12 crc kubenswrapper[4676]: I0124 00:27:12.657735 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8"] Jan 24 00:27:13 crc kubenswrapper[4676]: I0124 00:27:13.672981 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" event={"ID":"19b712ec-28a7-419f-9f09-7d0b0ecbf747","Type":"ContainerStarted","Data":"40f8aff558f949e11983226e96660908090ee1f149c1d729ce8f538212ca4716"} Jan 24 00:27:13 crc kubenswrapper[4676]: I0124 00:27:13.673353 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" event={"ID":"19b712ec-28a7-419f-9f09-7d0b0ecbf747","Type":"ContainerStarted","Data":"288f568b7f17ebcd72b74c000c5d8e959c55a88bbca9c451f6aac55cdca9fc0c"} Jan 24 00:27:13 crc kubenswrapper[4676]: I0124 00:27:13.717332 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" podStartSLOduration=2.2914809800000002 podStartE2EDuration="2.717307073s" podCreationTimestamp="2026-01-24 00:27:11 +0000 UTC" firstStartedPulling="2026-01-24 00:27:12.67122773 +0000 UTC m=+1416.701198731" lastFinishedPulling="2026-01-24 00:27:13.097053813 +0000 UTC m=+1417.127024824" observedRunningTime="2026-01-24 00:27:13.70457896 +0000 UTC m=+1417.734550001" watchObservedRunningTime="2026-01-24 00:27:13.717307073 +0000 UTC m=+1417.747278114" Jan 24 00:27:16 crc kubenswrapper[4676]: I0124 00:27:16.717326 4676 generic.go:334] "Generic (PLEG): container finished" podID="19b712ec-28a7-419f-9f09-7d0b0ecbf747" containerID="40f8aff558f949e11983226e96660908090ee1f149c1d729ce8f538212ca4716" exitCode=0 Jan 24 00:27:16 crc kubenswrapper[4676]: I0124 00:27:16.717437 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" event={"ID":"19b712ec-28a7-419f-9f09-7d0b0ecbf747","Type":"ContainerDied","Data":"40f8aff558f949e11983226e96660908090ee1f149c1d729ce8f538212ca4716"} Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.136351 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.272477 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-inventory\") pod \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.272920 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slkrn\" (UniqueName: \"kubernetes.io/projected/19b712ec-28a7-419f-9f09-7d0b0ecbf747-kube-api-access-slkrn\") pod \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.272969 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-ssh-key-openstack-edpm-ipam\") pod \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\" (UID: \"19b712ec-28a7-419f-9f09-7d0b0ecbf747\") " Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.292359 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19b712ec-28a7-419f-9f09-7d0b0ecbf747-kube-api-access-slkrn" (OuterVolumeSpecName: "kube-api-access-slkrn") pod "19b712ec-28a7-419f-9f09-7d0b0ecbf747" (UID: "19b712ec-28a7-419f-9f09-7d0b0ecbf747"). InnerVolumeSpecName "kube-api-access-slkrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.305867 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-inventory" (OuterVolumeSpecName: "inventory") pod "19b712ec-28a7-419f-9f09-7d0b0ecbf747" (UID: "19b712ec-28a7-419f-9f09-7d0b0ecbf747"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.315494 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "19b712ec-28a7-419f-9f09-7d0b0ecbf747" (UID: "19b712ec-28a7-419f-9f09-7d0b0ecbf747"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.375418 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.375520 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slkrn\" (UniqueName: \"kubernetes.io/projected/19b712ec-28a7-419f-9f09-7d0b0ecbf747-kube-api-access-slkrn\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.375573 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/19b712ec-28a7-419f-9f09-7d0b0ecbf747-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.742339 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" event={"ID":"19b712ec-28a7-419f-9f09-7d0b0ecbf747","Type":"ContainerDied","Data":"288f568b7f17ebcd72b74c000c5d8e959c55a88bbca9c451f6aac55cdca9fc0c"} Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.742444 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="288f568b7f17ebcd72b74c000c5d8e959c55a88bbca9c451f6aac55cdca9fc0c" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.742503 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qx6r8" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.826697 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp"] Jan 24 00:27:18 crc kubenswrapper[4676]: E0124 00:27:18.827085 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b712ec-28a7-419f-9f09-7d0b0ecbf747" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.827100 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b712ec-28a7-419f-9f09-7d0b0ecbf747" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.827278 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b712ec-28a7-419f-9f09-7d0b0ecbf747" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.827895 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.833263 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.833597 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.833710 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.833711 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.862227 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp"] Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.989940 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.990047 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cctcl\" (UniqueName: \"kubernetes.io/projected/1fd4e1f4-0772-493a-b929-6e93470f9abf-kube-api-access-cctcl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.990081 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:18 crc kubenswrapper[4676]: I0124 00:27:18.990153 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.091681 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.091986 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.092084 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cctcl\" (UniqueName: \"kubernetes.io/projected/1fd4e1f4-0772-493a-b929-6e93470f9abf-kube-api-access-cctcl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.092169 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.098296 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.098447 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.101416 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.114062 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cctcl\" (UniqueName: \"kubernetes.io/projected/1fd4e1f4-0772-493a-b929-6e93470f9abf-kube-api-access-cctcl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.147500 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:27:19 crc kubenswrapper[4676]: I0124 00:27:19.779667 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp"] Jan 24 00:27:20 crc kubenswrapper[4676]: I0124 00:27:20.763301 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" event={"ID":"1fd4e1f4-0772-493a-b929-6e93470f9abf","Type":"ContainerStarted","Data":"990efd15c2bf2faf18318ad60df1588709f1e3dfd780fd07d20dad8fa6ff65f7"} Jan 24 00:27:20 crc kubenswrapper[4676]: I0124 00:27:20.763754 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" event={"ID":"1fd4e1f4-0772-493a-b929-6e93470f9abf","Type":"ContainerStarted","Data":"b1076e008af5cd31c6b2d2ff588ee49398925df4037135eafcda7ef3f9b4a900"} Jan 24 00:27:20 crc kubenswrapper[4676]: I0124 00:27:20.787510 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" podStartSLOduration=2.295037624 podStartE2EDuration="2.787482454s" podCreationTimestamp="2026-01-24 00:27:18 +0000 UTC" firstStartedPulling="2026-01-24 00:27:19.793972178 +0000 UTC m=+1423.823943199" lastFinishedPulling="2026-01-24 00:27:20.286417008 +0000 UTC m=+1424.316388029" observedRunningTime="2026-01-24 00:27:20.782877763 +0000 UTC m=+1424.812848774" watchObservedRunningTime="2026-01-24 00:27:20.787482454 +0000 UTC m=+1424.817453455" Jan 24 00:27:40 crc kubenswrapper[4676]: I0124 00:27:40.062832 4676 scope.go:117] "RemoveContainer" containerID="d2ed0ffdbd6bb9c785556f969ad9aeabbea68ca6c96b8f79735c762793724bfe" Jan 24 00:27:40 crc kubenswrapper[4676]: I0124 00:27:40.104503 4676 scope.go:117] "RemoveContainer" containerID="f59e9569d48da0357305a5589f2cf90a01ccd73d4a805dd25787bb7bb6ebfba6" Jan 24 00:28:40 crc kubenswrapper[4676]: I0124 00:28:40.317710 4676 scope.go:117] "RemoveContainer" containerID="51869c6b7eeadd27a59a7b34889ba0093a949ad78c0444eec3606959ac5ba339" Jan 24 00:29:09 crc kubenswrapper[4676]: I0124 00:29:09.364418 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:29:09 crc kubenswrapper[4676]: I0124 00:29:09.365118 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:29:39 crc kubenswrapper[4676]: I0124 00:29:39.364447 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:29:39 crc kubenswrapper[4676]: I0124 00:29:39.365036 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:29:40 crc kubenswrapper[4676]: I0124 00:29:40.406103 4676 scope.go:117] "RemoveContainer" containerID="1da12014ab624fa95aa24ddc28c008acbe5db6c02d4d9e96a348452a0b9b9e38" Jan 24 00:29:40 crc kubenswrapper[4676]: I0124 00:29:40.443345 4676 scope.go:117] "RemoveContainer" containerID="3c1cdcc2ffd783555c7e0583ecb6ac91d2d0c15034969f7f5f6f8dba79f95867" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.188589 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj"] Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.191857 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.199955 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.200274 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.218178 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc94ad7e-9d49-4100-a653-8957ee663195-config-volume\") pod \"collect-profiles-29486910-w5nqj\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.218831 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzjs8\" (UniqueName: \"kubernetes.io/projected/fc94ad7e-9d49-4100-a653-8957ee663195-kube-api-access-vzjs8\") pod \"collect-profiles-29486910-w5nqj\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.218873 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc94ad7e-9d49-4100-a653-8957ee663195-secret-volume\") pod \"collect-profiles-29486910-w5nqj\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.229997 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj"] Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.320421 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc94ad7e-9d49-4100-a653-8957ee663195-config-volume\") pod \"collect-profiles-29486910-w5nqj\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.320550 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzjs8\" (UniqueName: \"kubernetes.io/projected/fc94ad7e-9d49-4100-a653-8957ee663195-kube-api-access-vzjs8\") pod \"collect-profiles-29486910-w5nqj\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.320577 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc94ad7e-9d49-4100-a653-8957ee663195-secret-volume\") pod \"collect-profiles-29486910-w5nqj\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.321498 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc94ad7e-9d49-4100-a653-8957ee663195-config-volume\") pod \"collect-profiles-29486910-w5nqj\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.335005 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc94ad7e-9d49-4100-a653-8957ee663195-secret-volume\") pod \"collect-profiles-29486910-w5nqj\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.340413 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzjs8\" (UniqueName: \"kubernetes.io/projected/fc94ad7e-9d49-4100-a653-8957ee663195-kube-api-access-vzjs8\") pod \"collect-profiles-29486910-w5nqj\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:00 crc kubenswrapper[4676]: I0124 00:30:00.528557 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.023084 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj"] Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.521364 4676 generic.go:334] "Generic (PLEG): container finished" podID="fc94ad7e-9d49-4100-a653-8957ee663195" containerID="8db303a059dedc6aec4dda92c8f1bdb2072d4f7fcc3f6c35461f6c01badf580c" exitCode=0 Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.521462 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" event={"ID":"fc94ad7e-9d49-4100-a653-8957ee663195","Type":"ContainerDied","Data":"8db303a059dedc6aec4dda92c8f1bdb2072d4f7fcc3f6c35461f6c01badf580c"} Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.521717 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" event={"ID":"fc94ad7e-9d49-4100-a653-8957ee663195","Type":"ContainerStarted","Data":"2baa3f0128f1f45531ea208204732c90b4b655af648d15f19a28ea9c106d6c67"} Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.615956 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-69zb6"] Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.618093 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.631673 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-69zb6"] Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.752680 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gxkn\" (UniqueName: \"kubernetes.io/projected/6f07f20b-f218-4f5b-9f8a-127ca3b44402-kube-api-access-5gxkn\") pod \"redhat-marketplace-69zb6\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.752947 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-utilities\") pod \"redhat-marketplace-69zb6\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.753109 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-catalog-content\") pod \"redhat-marketplace-69zb6\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.854427 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-catalog-content\") pod \"redhat-marketplace-69zb6\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.854649 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gxkn\" (UniqueName: \"kubernetes.io/projected/6f07f20b-f218-4f5b-9f8a-127ca3b44402-kube-api-access-5gxkn\") pod \"redhat-marketplace-69zb6\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.854763 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-utilities\") pod \"redhat-marketplace-69zb6\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.855038 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-catalog-content\") pod \"redhat-marketplace-69zb6\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.855511 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-utilities\") pod \"redhat-marketplace-69zb6\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.885153 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gxkn\" (UniqueName: \"kubernetes.io/projected/6f07f20b-f218-4f5b-9f8a-127ca3b44402-kube-api-access-5gxkn\") pod \"redhat-marketplace-69zb6\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:01 crc kubenswrapper[4676]: I0124 00:30:01.951759 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:02 crc kubenswrapper[4676]: I0124 00:30:02.450739 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-69zb6"] Jan 24 00:30:02 crc kubenswrapper[4676]: I0124 00:30:02.534116 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69zb6" event={"ID":"6f07f20b-f218-4f5b-9f8a-127ca3b44402","Type":"ContainerStarted","Data":"fb5a366881b73b432a758371a8751ca97872bde3cb9d50974c2e5c2b309a7330"} Jan 24 00:30:02 crc kubenswrapper[4676]: I0124 00:30:02.878554 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:02 crc kubenswrapper[4676]: I0124 00:30:02.975570 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzjs8\" (UniqueName: \"kubernetes.io/projected/fc94ad7e-9d49-4100-a653-8957ee663195-kube-api-access-vzjs8\") pod \"fc94ad7e-9d49-4100-a653-8957ee663195\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " Jan 24 00:30:02 crc kubenswrapper[4676]: I0124 00:30:02.975667 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc94ad7e-9d49-4100-a653-8957ee663195-secret-volume\") pod \"fc94ad7e-9d49-4100-a653-8957ee663195\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " Jan 24 00:30:02 crc kubenswrapper[4676]: I0124 00:30:02.975812 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc94ad7e-9d49-4100-a653-8957ee663195-config-volume\") pod \"fc94ad7e-9d49-4100-a653-8957ee663195\" (UID: \"fc94ad7e-9d49-4100-a653-8957ee663195\") " Jan 24 00:30:02 crc kubenswrapper[4676]: I0124 00:30:02.977169 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc94ad7e-9d49-4100-a653-8957ee663195-config-volume" (OuterVolumeSpecName: "config-volume") pod "fc94ad7e-9d49-4100-a653-8957ee663195" (UID: "fc94ad7e-9d49-4100-a653-8957ee663195"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:30:02 crc kubenswrapper[4676]: I0124 00:30:02.984814 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc94ad7e-9d49-4100-a653-8957ee663195-kube-api-access-vzjs8" (OuterVolumeSpecName: "kube-api-access-vzjs8") pod "fc94ad7e-9d49-4100-a653-8957ee663195" (UID: "fc94ad7e-9d49-4100-a653-8957ee663195"). InnerVolumeSpecName "kube-api-access-vzjs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:30:02 crc kubenswrapper[4676]: I0124 00:30:02.990493 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc94ad7e-9d49-4100-a653-8957ee663195-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fc94ad7e-9d49-4100-a653-8957ee663195" (UID: "fc94ad7e-9d49-4100-a653-8957ee663195"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:30:03 crc kubenswrapper[4676]: I0124 00:30:03.078884 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzjs8\" (UniqueName: \"kubernetes.io/projected/fc94ad7e-9d49-4100-a653-8957ee663195-kube-api-access-vzjs8\") on node \"crc\" DevicePath \"\"" Jan 24 00:30:03 crc kubenswrapper[4676]: I0124 00:30:03.078936 4676 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fc94ad7e-9d49-4100-a653-8957ee663195-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 24 00:30:03 crc kubenswrapper[4676]: I0124 00:30:03.078960 4676 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fc94ad7e-9d49-4100-a653-8957ee663195-config-volume\") on node \"crc\" DevicePath \"\"" Jan 24 00:30:03 crc kubenswrapper[4676]: I0124 00:30:03.550662 4676 generic.go:334] "Generic (PLEG): container finished" podID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerID="1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700" exitCode=0 Jan 24 00:30:03 crc kubenswrapper[4676]: I0124 00:30:03.550739 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69zb6" event={"ID":"6f07f20b-f218-4f5b-9f8a-127ca3b44402","Type":"ContainerDied","Data":"1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700"} Jan 24 00:30:03 crc kubenswrapper[4676]: I0124 00:30:03.553075 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" event={"ID":"fc94ad7e-9d49-4100-a653-8957ee663195","Type":"ContainerDied","Data":"2baa3f0128f1f45531ea208204732c90b4b655af648d15f19a28ea9c106d6c67"} Jan 24 00:30:03 crc kubenswrapper[4676]: I0124 00:30:03.553096 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2baa3f0128f1f45531ea208204732c90b4b655af648d15f19a28ea9c106d6c67" Jan 24 00:30:03 crc kubenswrapper[4676]: I0124 00:30:03.553151 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486910-w5nqj" Jan 24 00:30:03 crc kubenswrapper[4676]: I0124 00:30:03.553439 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:30:05 crc kubenswrapper[4676]: I0124 00:30:05.580244 4676 generic.go:334] "Generic (PLEG): container finished" podID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerID="1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f" exitCode=0 Jan 24 00:30:05 crc kubenswrapper[4676]: I0124 00:30:05.580359 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69zb6" event={"ID":"6f07f20b-f218-4f5b-9f8a-127ca3b44402","Type":"ContainerDied","Data":"1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f"} Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.198415 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rmln4"] Jan 24 00:30:06 crc kubenswrapper[4676]: E0124 00:30:06.199153 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc94ad7e-9d49-4100-a653-8957ee663195" containerName="collect-profiles" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.199173 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc94ad7e-9d49-4100-a653-8957ee663195" containerName="collect-profiles" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.199358 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc94ad7e-9d49-4100-a653-8957ee663195" containerName="collect-profiles" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.201050 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.225733 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rmln4"] Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.273800 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-catalog-content\") pod \"certified-operators-rmln4\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.273882 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-utilities\") pod \"certified-operators-rmln4\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.273986 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg6vm\" (UniqueName: \"kubernetes.io/projected/df68cc49-9d8a-4c50-a379-35a2d56491e7-kube-api-access-mg6vm\") pod \"certified-operators-rmln4\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.376914 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-catalog-content\") pod \"certified-operators-rmln4\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.377344 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-utilities\") pod \"certified-operators-rmln4\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.377559 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg6vm\" (UniqueName: \"kubernetes.io/projected/df68cc49-9d8a-4c50-a379-35a2d56491e7-kube-api-access-mg6vm\") pod \"certified-operators-rmln4\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.377686 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-catalog-content\") pod \"certified-operators-rmln4\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.378821 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-utilities\") pod \"certified-operators-rmln4\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.398008 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg6vm\" (UniqueName: \"kubernetes.io/projected/df68cc49-9d8a-4c50-a379-35a2d56491e7-kube-api-access-mg6vm\") pod \"certified-operators-rmln4\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.586351 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.591242 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69zb6" event={"ID":"6f07f20b-f218-4f5b-9f8a-127ca3b44402","Type":"ContainerStarted","Data":"79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3"} Jan 24 00:30:06 crc kubenswrapper[4676]: I0124 00:30:06.626279 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-69zb6" podStartSLOduration=3.128961409 podStartE2EDuration="5.626258238s" podCreationTimestamp="2026-01-24 00:30:01 +0000 UTC" firstStartedPulling="2026-01-24 00:30:03.553045913 +0000 UTC m=+1587.583016944" lastFinishedPulling="2026-01-24 00:30:06.050342762 +0000 UTC m=+1590.080313773" observedRunningTime="2026-01-24 00:30:06.616699026 +0000 UTC m=+1590.646670037" watchObservedRunningTime="2026-01-24 00:30:06.626258238 +0000 UTC m=+1590.656229239" Jan 24 00:30:07 crc kubenswrapper[4676]: I0124 00:30:07.141283 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rmln4"] Jan 24 00:30:07 crc kubenswrapper[4676]: I0124 00:30:07.628307 4676 generic.go:334] "Generic (PLEG): container finished" podID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerID="2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2" exitCode=0 Jan 24 00:30:07 crc kubenswrapper[4676]: I0124 00:30:07.630195 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmln4" event={"ID":"df68cc49-9d8a-4c50-a379-35a2d56491e7","Type":"ContainerDied","Data":"2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2"} Jan 24 00:30:07 crc kubenswrapper[4676]: I0124 00:30:07.630818 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmln4" event={"ID":"df68cc49-9d8a-4c50-a379-35a2d56491e7","Type":"ContainerStarted","Data":"de097375a51280c39d2e2ca9ddfc9fe7339a04c32101fb8241605b662c33ab05"} Jan 24 00:30:08 crc kubenswrapper[4676]: I0124 00:30:08.640609 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmln4" event={"ID":"df68cc49-9d8a-4c50-a379-35a2d56491e7","Type":"ContainerStarted","Data":"368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792"} Jan 24 00:30:09 crc kubenswrapper[4676]: I0124 00:30:09.364582 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:30:09 crc kubenswrapper[4676]: I0124 00:30:09.364673 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:30:09 crc kubenswrapper[4676]: I0124 00:30:09.364743 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:30:09 crc kubenswrapper[4676]: I0124 00:30:09.365967 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:30:09 crc kubenswrapper[4676]: I0124 00:30:09.366082 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" gracePeriod=600 Jan 24 00:30:11 crc kubenswrapper[4676]: I0124 00:30:11.953291 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:11 crc kubenswrapper[4676]: I0124 00:30:11.953972 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:12 crc kubenswrapper[4676]: I0124 00:30:12.015289 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:12 crc kubenswrapper[4676]: I0124 00:30:12.678756 4676 generic.go:334] "Generic (PLEG): container finished" podID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerID="368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792" exitCode=0 Jan 24 00:30:12 crc kubenswrapper[4676]: I0124 00:30:12.678824 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmln4" event={"ID":"df68cc49-9d8a-4c50-a379-35a2d56491e7","Type":"ContainerDied","Data":"368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792"} Jan 24 00:30:12 crc kubenswrapper[4676]: I0124 00:30:12.683685 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" exitCode=0 Jan 24 00:30:12 crc kubenswrapper[4676]: I0124 00:30:12.683790 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751"} Jan 24 00:30:12 crc kubenswrapper[4676]: I0124 00:30:12.683864 4676 scope.go:117] "RemoveContainer" containerID="6c0fdc4fa29c1a85e06a8e0b7d6899a3299af1dbe5c0f08e67bee66057d6c55e" Jan 24 00:30:12 crc kubenswrapper[4676]: I0124 00:30:12.742163 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:13 crc kubenswrapper[4676]: E0124 00:30:13.464981 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:30:13 crc kubenswrapper[4676]: I0124 00:30:13.700742 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:30:13 crc kubenswrapper[4676]: E0124 00:30:13.701227 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:30:14 crc kubenswrapper[4676]: I0124 00:30:14.190846 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-69zb6"] Jan 24 00:30:14 crc kubenswrapper[4676]: I0124 00:30:14.711416 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-69zb6" podUID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerName="registry-server" containerID="cri-o://79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3" gracePeriod=2 Jan 24 00:30:14 crc kubenswrapper[4676]: I0124 00:30:14.711759 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmln4" event={"ID":"df68cc49-9d8a-4c50-a379-35a2d56491e7","Type":"ContainerStarted","Data":"e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa"} Jan 24 00:30:14 crc kubenswrapper[4676]: I0124 00:30:14.739637 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rmln4" podStartSLOduration=2.284833619 podStartE2EDuration="8.73961936s" podCreationTimestamp="2026-01-24 00:30:06 +0000 UTC" firstStartedPulling="2026-01-24 00:30:07.631957521 +0000 UTC m=+1591.661928522" lastFinishedPulling="2026-01-24 00:30:14.086743262 +0000 UTC m=+1598.116714263" observedRunningTime="2026-01-24 00:30:14.738695724 +0000 UTC m=+1598.768666735" watchObservedRunningTime="2026-01-24 00:30:14.73961936 +0000 UTC m=+1598.769590361" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.286693 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.356479 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-catalog-content\") pod \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.356722 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gxkn\" (UniqueName: \"kubernetes.io/projected/6f07f20b-f218-4f5b-9f8a-127ca3b44402-kube-api-access-5gxkn\") pod \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.356894 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-utilities\") pod \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\" (UID: \"6f07f20b-f218-4f5b-9f8a-127ca3b44402\") " Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.374493 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-utilities" (OuterVolumeSpecName: "utilities") pod "6f07f20b-f218-4f5b-9f8a-127ca3b44402" (UID: "6f07f20b-f218-4f5b-9f8a-127ca3b44402"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.379550 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f07f20b-f218-4f5b-9f8a-127ca3b44402-kube-api-access-5gxkn" (OuterVolumeSpecName: "kube-api-access-5gxkn") pod "6f07f20b-f218-4f5b-9f8a-127ca3b44402" (UID: "6f07f20b-f218-4f5b-9f8a-127ca3b44402"). InnerVolumeSpecName "kube-api-access-5gxkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.383178 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6f07f20b-f218-4f5b-9f8a-127ca3b44402" (UID: "6f07f20b-f218-4f5b-9f8a-127ca3b44402"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.458524 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gxkn\" (UniqueName: \"kubernetes.io/projected/6f07f20b-f218-4f5b-9f8a-127ca3b44402-kube-api-access-5gxkn\") on node \"crc\" DevicePath \"\"" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.458555 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.458564 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6f07f20b-f218-4f5b-9f8a-127ca3b44402-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.724063 4676 generic.go:334] "Generic (PLEG): container finished" podID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerID="79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3" exitCode=0 Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.724121 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69zb6" event={"ID":"6f07f20b-f218-4f5b-9f8a-127ca3b44402","Type":"ContainerDied","Data":"79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3"} Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.724163 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69zb6" event={"ID":"6f07f20b-f218-4f5b-9f8a-127ca3b44402","Type":"ContainerDied","Data":"fb5a366881b73b432a758371a8751ca97872bde3cb9d50974c2e5c2b309a7330"} Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.724194 4676 scope.go:117] "RemoveContainer" containerID="79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.724365 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69zb6" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.746512 4676 scope.go:117] "RemoveContainer" containerID="1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.791082 4676 scope.go:117] "RemoveContainer" containerID="1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.797963 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-69zb6"] Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.819793 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-69zb6"] Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.834002 4676 scope.go:117] "RemoveContainer" containerID="79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3" Jan 24 00:30:15 crc kubenswrapper[4676]: E0124 00:30:15.834343 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3\": container with ID starting with 79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3 not found: ID does not exist" containerID="79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.834387 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3"} err="failed to get container status \"79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3\": rpc error: code = NotFound desc = could not find container \"79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3\": container with ID starting with 79c7387ea68d950659fb07676ebcecdc5c95a40f02a96f9a9d61a62fac3d6be3 not found: ID does not exist" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.834413 4676 scope.go:117] "RemoveContainer" containerID="1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f" Jan 24 00:30:15 crc kubenswrapper[4676]: E0124 00:30:15.834609 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f\": container with ID starting with 1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f not found: ID does not exist" containerID="1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.834630 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f"} err="failed to get container status \"1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f\": rpc error: code = NotFound desc = could not find container \"1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f\": container with ID starting with 1f4c7897af2181f99f84074c3f9548df41f7014f5581145d277c051dea39a12f not found: ID does not exist" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.834645 4676 scope.go:117] "RemoveContainer" containerID="1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700" Jan 24 00:30:15 crc kubenswrapper[4676]: E0124 00:30:15.834812 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700\": container with ID starting with 1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700 not found: ID does not exist" containerID="1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700" Jan 24 00:30:15 crc kubenswrapper[4676]: I0124 00:30:15.834837 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700"} err="failed to get container status \"1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700\": rpc error: code = NotFound desc = could not find container \"1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700\": container with ID starting with 1df2094ba92d96207746ec06cc7516275de92ffeb2ec963064e4a5ed06a53700 not found: ID does not exist" Jan 24 00:30:16 crc kubenswrapper[4676]: I0124 00:30:16.269687 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" path="/var/lib/kubelet/pods/6f07f20b-f218-4f5b-9f8a-127ca3b44402/volumes" Jan 24 00:30:16 crc kubenswrapper[4676]: I0124 00:30:16.587703 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:16 crc kubenswrapper[4676]: I0124 00:30:16.588150 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:16 crc kubenswrapper[4676]: I0124 00:30:16.646933 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:24 crc kubenswrapper[4676]: I0124 00:30:24.256077 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:30:24 crc kubenswrapper[4676]: E0124 00:30:24.257284 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:30:26 crc kubenswrapper[4676]: I0124 00:30:26.647798 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:26 crc kubenswrapper[4676]: I0124 00:30:26.699031 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rmln4"] Jan 24 00:30:26 crc kubenswrapper[4676]: I0124 00:30:26.861953 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rmln4" podUID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerName="registry-server" containerID="cri-o://e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa" gracePeriod=2 Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.307929 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.395703 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-utilities\") pod \"df68cc49-9d8a-4c50-a379-35a2d56491e7\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.395779 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg6vm\" (UniqueName: \"kubernetes.io/projected/df68cc49-9d8a-4c50-a379-35a2d56491e7-kube-api-access-mg6vm\") pod \"df68cc49-9d8a-4c50-a379-35a2d56491e7\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.395864 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-catalog-content\") pod \"df68cc49-9d8a-4c50-a379-35a2d56491e7\" (UID: \"df68cc49-9d8a-4c50-a379-35a2d56491e7\") " Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.396552 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-utilities" (OuterVolumeSpecName: "utilities") pod "df68cc49-9d8a-4c50-a379-35a2d56491e7" (UID: "df68cc49-9d8a-4c50-a379-35a2d56491e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.403049 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.404655 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df68cc49-9d8a-4c50-a379-35a2d56491e7-kube-api-access-mg6vm" (OuterVolumeSpecName: "kube-api-access-mg6vm") pod "df68cc49-9d8a-4c50-a379-35a2d56491e7" (UID: "df68cc49-9d8a-4c50-a379-35a2d56491e7"). InnerVolumeSpecName "kube-api-access-mg6vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.452649 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df68cc49-9d8a-4c50-a379-35a2d56491e7" (UID: "df68cc49-9d8a-4c50-a379-35a2d56491e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.504334 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df68cc49-9d8a-4c50-a379-35a2d56491e7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.504363 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg6vm\" (UniqueName: \"kubernetes.io/projected/df68cc49-9d8a-4c50-a379-35a2d56491e7-kube-api-access-mg6vm\") on node \"crc\" DevicePath \"\"" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.876929 4676 generic.go:334] "Generic (PLEG): container finished" podID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerID="e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa" exitCode=0 Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.877019 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmln4" event={"ID":"df68cc49-9d8a-4c50-a379-35a2d56491e7","Type":"ContainerDied","Data":"e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa"} Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.877047 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rmln4" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.877292 4676 scope.go:117] "RemoveContainer" containerID="e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.877260 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rmln4" event={"ID":"df68cc49-9d8a-4c50-a379-35a2d56491e7","Type":"ContainerDied","Data":"de097375a51280c39d2e2ca9ddfc9fe7339a04c32101fb8241605b662c33ab05"} Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.898511 4676 scope.go:117] "RemoveContainer" containerID="368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.927532 4676 scope.go:117] "RemoveContainer" containerID="2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.941553 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rmln4"] Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.953500 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rmln4"] Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.983275 4676 scope.go:117] "RemoveContainer" containerID="e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa" Jan 24 00:30:27 crc kubenswrapper[4676]: E0124 00:30:27.983907 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa\": container with ID starting with e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa not found: ID does not exist" containerID="e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.983957 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa"} err="failed to get container status \"e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa\": rpc error: code = NotFound desc = could not find container \"e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa\": container with ID starting with e82244cf7f7fdf513932fe8199f8fe13f14ccd1a9b143f6c05a4a817e6564dfa not found: ID does not exist" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.983990 4676 scope.go:117] "RemoveContainer" containerID="368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792" Jan 24 00:30:27 crc kubenswrapper[4676]: E0124 00:30:27.984755 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792\": container with ID starting with 368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792 not found: ID does not exist" containerID="368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.984786 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792"} err="failed to get container status \"368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792\": rpc error: code = NotFound desc = could not find container \"368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792\": container with ID starting with 368d2b86f772740717cfdf15a60a1fd8f7b0f9d3fd68fb9aa7cef21ebb376792 not found: ID does not exist" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.984809 4676 scope.go:117] "RemoveContainer" containerID="2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2" Jan 24 00:30:27 crc kubenswrapper[4676]: E0124 00:30:27.985093 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2\": container with ID starting with 2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2 not found: ID does not exist" containerID="2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2" Jan 24 00:30:27 crc kubenswrapper[4676]: I0124 00:30:27.985124 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2"} err="failed to get container status \"2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2\": rpc error: code = NotFound desc = could not find container \"2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2\": container with ID starting with 2b00bdedbfb520a4d73d3abd4a42b5761963be56c158ebdc25278c30b6bd67d2 not found: ID does not exist" Jan 24 00:30:28 crc kubenswrapper[4676]: I0124 00:30:28.266612 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df68cc49-9d8a-4c50-a379-35a2d56491e7" path="/var/lib/kubelet/pods/df68cc49-9d8a-4c50-a379-35a2d56491e7/volumes" Jan 24 00:30:37 crc kubenswrapper[4676]: I0124 00:30:37.255074 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:30:37 crc kubenswrapper[4676]: E0124 00:30:37.257131 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:30:40 crc kubenswrapper[4676]: I0124 00:30:40.533982 4676 scope.go:117] "RemoveContainer" containerID="b50dc415c434720b0458f8bff74066a1f5105e74beeb732a3d51cdf130cbe5d3" Jan 24 00:30:40 crc kubenswrapper[4676]: I0124 00:30:40.568318 4676 scope.go:117] "RemoveContainer" containerID="6301d47d4294644c70fd618b97e56c019270427fbdab822047ac3bfe88e7e594" Jan 24 00:30:52 crc kubenswrapper[4676]: I0124 00:30:52.256883 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:30:52 crc kubenswrapper[4676]: E0124 00:30:52.257695 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:30:58 crc kubenswrapper[4676]: I0124 00:30:58.078978 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-ph986"] Jan 24 00:30:58 crc kubenswrapper[4676]: I0124 00:30:58.094435 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-d7f6-account-create-update-brcr8"] Jan 24 00:30:58 crc kubenswrapper[4676]: I0124 00:30:58.103790 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-ph986"] Jan 24 00:30:58 crc kubenswrapper[4676]: I0124 00:30:58.112181 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-d7f6-account-create-update-brcr8"] Jan 24 00:30:58 crc kubenswrapper[4676]: I0124 00:30:58.273797 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925957f4-9ca1-4e46-b3ac-0dd0ca5713cd" path="/var/lib/kubelet/pods/925957f4-9ca1-4e46-b3ac-0dd0ca5713cd/volumes" Jan 24 00:30:58 crc kubenswrapper[4676]: I0124 00:30:58.275540 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98c25d86-7dc6-4b15-9511-3a7cbf4c7592" path="/var/lib/kubelet/pods/98c25d86-7dc6-4b15-9511-3a7cbf4c7592/volumes" Jan 24 00:31:02 crc kubenswrapper[4676]: I0124 00:31:02.058540 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-x4ptm"] Jan 24 00:31:02 crc kubenswrapper[4676]: I0124 00:31:02.077277 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-60c3-account-create-update-xz4kb"] Jan 24 00:31:02 crc kubenswrapper[4676]: I0124 00:31:02.090754 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-x4ptm"] Jan 24 00:31:02 crc kubenswrapper[4676]: I0124 00:31:02.104557 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-60c3-account-create-update-xz4kb"] Jan 24 00:31:02 crc kubenswrapper[4676]: I0124 00:31:02.272295 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42" path="/var/lib/kubelet/pods/4ebbdd8e-9c69-4d2a-8f14-14bc152c1a42/volumes" Jan 24 00:31:02 crc kubenswrapper[4676]: I0124 00:31:02.274270 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6d8987f-6f37-4773-99ad-6bc7fd8971f2" path="/var/lib/kubelet/pods/d6d8987f-6f37-4773-99ad-6bc7fd8971f2/volumes" Jan 24 00:31:03 crc kubenswrapper[4676]: I0124 00:31:03.043233 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-prlkx"] Jan 24 00:31:03 crc kubenswrapper[4676]: I0124 00:31:03.055669 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-prlkx"] Jan 24 00:31:03 crc kubenswrapper[4676]: I0124 00:31:03.067556 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-278e-account-create-update-8fwkd"] Jan 24 00:31:03 crc kubenswrapper[4676]: I0124 00:31:03.074876 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-278e-account-create-update-8fwkd"] Jan 24 00:31:03 crc kubenswrapper[4676]: I0124 00:31:03.219120 4676 generic.go:334] "Generic (PLEG): container finished" podID="1fd4e1f4-0772-493a-b929-6e93470f9abf" containerID="990efd15c2bf2faf18318ad60df1588709f1e3dfd780fd07d20dad8fa6ff65f7" exitCode=0 Jan 24 00:31:03 crc kubenswrapper[4676]: I0124 00:31:03.219683 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" event={"ID":"1fd4e1f4-0772-493a-b929-6e93470f9abf","Type":"ContainerDied","Data":"990efd15c2bf2faf18318ad60df1588709f1e3dfd780fd07d20dad8fa6ff65f7"} Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.279457 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2374740d-d0b0-4a0c-b23e-d41d3ad524ac" path="/var/lib/kubelet/pods/2374740d-d0b0-4a0c-b23e-d41d3ad524ac/volumes" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.281540 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="439dce27-2a59-49a2-beab-82482c1ce3cb" path="/var/lib/kubelet/pods/439dce27-2a59-49a2-beab-82482c1ce3cb/volumes" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.655480 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.760328 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cctcl\" (UniqueName: \"kubernetes.io/projected/1fd4e1f4-0772-493a-b929-6e93470f9abf-kube-api-access-cctcl\") pod \"1fd4e1f4-0772-493a-b929-6e93470f9abf\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.760397 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-bootstrap-combined-ca-bundle\") pod \"1fd4e1f4-0772-493a-b929-6e93470f9abf\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.760497 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-ssh-key-openstack-edpm-ipam\") pod \"1fd4e1f4-0772-493a-b929-6e93470f9abf\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.760538 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-inventory\") pod \"1fd4e1f4-0772-493a-b929-6e93470f9abf\" (UID: \"1fd4e1f4-0772-493a-b929-6e93470f9abf\") " Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.767118 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1fd4e1f4-0772-493a-b929-6e93470f9abf" (UID: "1fd4e1f4-0772-493a-b929-6e93470f9abf"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.770542 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd4e1f4-0772-493a-b929-6e93470f9abf-kube-api-access-cctcl" (OuterVolumeSpecName: "kube-api-access-cctcl") pod "1fd4e1f4-0772-493a-b929-6e93470f9abf" (UID: "1fd4e1f4-0772-493a-b929-6e93470f9abf"). InnerVolumeSpecName "kube-api-access-cctcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.796157 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1fd4e1f4-0772-493a-b929-6e93470f9abf" (UID: "1fd4e1f4-0772-493a-b929-6e93470f9abf"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.798557 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-inventory" (OuterVolumeSpecName: "inventory") pod "1fd4e1f4-0772-493a-b929-6e93470f9abf" (UID: "1fd4e1f4-0772-493a-b929-6e93470f9abf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.863674 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.863986 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.863997 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cctcl\" (UniqueName: \"kubernetes.io/projected/1fd4e1f4-0772-493a-b929-6e93470f9abf-kube-api-access-cctcl\") on node \"crc\" DevicePath \"\"" Jan 24 00:31:04 crc kubenswrapper[4676]: I0124 00:31:04.864008 4676 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fd4e1f4-0772-493a-b929-6e93470f9abf-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.242334 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" event={"ID":"1fd4e1f4-0772-493a-b929-6e93470f9abf","Type":"ContainerDied","Data":"b1076e008af5cd31c6b2d2ff588ee49398925df4037135eafcda7ef3f9b4a900"} Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.242403 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1076e008af5cd31c6b2d2ff588ee49398925df4037135eafcda7ef3f9b4a900" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.242432 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.354248 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k"] Jan 24 00:31:05 crc kubenswrapper[4676]: E0124 00:31:05.354649 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerName="registry-server" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.354662 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerName="registry-server" Jan 24 00:31:05 crc kubenswrapper[4676]: E0124 00:31:05.354672 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerName="extract-utilities" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.354678 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerName="extract-utilities" Jan 24 00:31:05 crc kubenswrapper[4676]: E0124 00:31:05.354689 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerName="registry-server" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.354695 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerName="registry-server" Jan 24 00:31:05 crc kubenswrapper[4676]: E0124 00:31:05.354703 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd4e1f4-0772-493a-b929-6e93470f9abf" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.354709 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd4e1f4-0772-493a-b929-6e93470f9abf" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 24 00:31:05 crc kubenswrapper[4676]: E0124 00:31:05.354723 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerName="extract-utilities" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.354731 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerName="extract-utilities" Jan 24 00:31:05 crc kubenswrapper[4676]: E0124 00:31:05.354750 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerName="extract-content" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.354758 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerName="extract-content" Jan 24 00:31:05 crc kubenswrapper[4676]: E0124 00:31:05.354784 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerName="extract-content" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.354792 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerName="extract-content" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.354980 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="df68cc49-9d8a-4c50-a379-35a2d56491e7" containerName="registry-server" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.355002 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f07f20b-f218-4f5b-9f8a-127ca3b44402" containerName="registry-server" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.355019 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd4e1f4-0772-493a-b929-6e93470f9abf" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.355766 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.365186 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.365412 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.365559 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.365798 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.373166 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k"] Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.477207 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-skh8k\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.477273 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-skh8k\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.477310 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22ck7\" (UniqueName: \"kubernetes.io/projected/7e20d58a-5f01-4c23-9ab0-650d3ae76844-kube-api-access-22ck7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-skh8k\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.578557 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-skh8k\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.578606 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-skh8k\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.578645 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22ck7\" (UniqueName: \"kubernetes.io/projected/7e20d58a-5f01-4c23-9ab0-650d3ae76844-kube-api-access-22ck7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-skh8k\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.587477 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-skh8k\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.587956 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-skh8k\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.601001 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22ck7\" (UniqueName: \"kubernetes.io/projected/7e20d58a-5f01-4c23-9ab0-650d3ae76844-kube-api-access-22ck7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-skh8k\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:05 crc kubenswrapper[4676]: I0124 00:31:05.683504 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:31:06 crc kubenswrapper[4676]: I0124 00:31:06.080300 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k"] Jan 24 00:31:06 crc kubenswrapper[4676]: I0124 00:31:06.254860 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" event={"ID":"7e20d58a-5f01-4c23-9ab0-650d3ae76844","Type":"ContainerStarted","Data":"d5ce814e830d10bf24b9a1a0312e8af78b1ad869f8900e1ffd1c55e849d5a89e"} Jan 24 00:31:06 crc kubenswrapper[4676]: I0124 00:31:06.262810 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:31:06 crc kubenswrapper[4676]: E0124 00:31:06.263162 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:31:07 crc kubenswrapper[4676]: I0124 00:31:07.268245 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" event={"ID":"7e20d58a-5f01-4c23-9ab0-650d3ae76844","Type":"ContainerStarted","Data":"a00ce8325a13556bd7fc250572ccaf4efe71b1b28dfcc05e56a4b357eae4bb8d"} Jan 24 00:31:19 crc kubenswrapper[4676]: I0124 00:31:19.039442 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" podStartSLOduration=13.478746551 podStartE2EDuration="14.039422559s" podCreationTimestamp="2026-01-24 00:31:05 +0000 UTC" firstStartedPulling="2026-01-24 00:31:06.094777178 +0000 UTC m=+1650.124748219" lastFinishedPulling="2026-01-24 00:31:06.655453226 +0000 UTC m=+1650.685424227" observedRunningTime="2026-01-24 00:31:07.290710578 +0000 UTC m=+1651.320681579" watchObservedRunningTime="2026-01-24 00:31:19.039422559 +0000 UTC m=+1663.069393580" Jan 24 00:31:19 crc kubenswrapper[4676]: I0124 00:31:19.041135 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-drjg7"] Jan 24 00:31:19 crc kubenswrapper[4676]: I0124 00:31:19.050732 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-drjg7"] Jan 24 00:31:19 crc kubenswrapper[4676]: I0124 00:31:19.255845 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:31:19 crc kubenswrapper[4676]: E0124 00:31:19.256059 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:31:20 crc kubenswrapper[4676]: I0124 00:31:20.266172 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20421570-a66e-4903-937f-af2ab847b5f1" path="/var/lib/kubelet/pods/20421570-a66e-4903-937f-af2ab847b5f1/volumes" Jan 24 00:31:28 crc kubenswrapper[4676]: I0124 00:31:28.043669 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-kbtsc"] Jan 24 00:31:28 crc kubenswrapper[4676]: I0124 00:31:28.063499 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-kbtsc"] Jan 24 00:31:28 crc kubenswrapper[4676]: I0124 00:31:28.273858 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ebdc261-7c0d-49a2-9233-868fda906788" path="/var/lib/kubelet/pods/7ebdc261-7c0d-49a2-9233-868fda906788/volumes" Jan 24 00:31:30 crc kubenswrapper[4676]: I0124 00:31:30.256574 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:31:30 crc kubenswrapper[4676]: E0124 00:31:30.260681 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.036068 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-bcttg"] Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.044706 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-bcttg"] Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.267938 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b488c3f8-bc04-4c8d-98b7-72145ae9e948" path="/var/lib/kubelet/pods/b488c3f8-bc04-4c8d-98b7-72145ae9e948/volumes" Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.692494 4676 scope.go:117] "RemoveContainer" containerID="1c08827f5b622dc0d2e57d1709fbff0d00bb4f69592994ac4ce101a373fb04cb" Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.784115 4676 scope.go:117] "RemoveContainer" containerID="b1e8f656f78cf6d79b93f73d693ca8f011b12dbd49906ddf8afe723c5036938e" Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.807831 4676 scope.go:117] "RemoveContainer" containerID="d5c280bb67bfb03306230881277a4f3cf73d70180a6643cdcdc6c29c26f5cdcb" Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.862466 4676 scope.go:117] "RemoveContainer" containerID="89ec371006cc8981cf4d3c50f26065f195057a546aba2faa19252efcd286b309" Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.897763 4676 scope.go:117] "RemoveContainer" containerID="a55ff63829260fcafaf5bd727871713740d9ed36c94c4fc218cf31c8bb4fa53b" Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.938841 4676 scope.go:117] "RemoveContainer" containerID="7b1df20731b13a3b81d6ecd68e39c0140b372b18913afe443cbe4c6b00f28ea4" Jan 24 00:31:40 crc kubenswrapper[4676]: I0124 00:31:40.982990 4676 scope.go:117] "RemoveContainer" containerID="8280acdb35983168359f77fef7c062161338aa47a5bf092b620c75a0dd83736f" Jan 24 00:31:41 crc kubenswrapper[4676]: I0124 00:31:41.016689 4676 scope.go:117] "RemoveContainer" containerID="a94f61fa91f9d6c0d5e0b0e07bb624e5cb428e8002a06af7cee89d3dcb90c4cd" Jan 24 00:31:41 crc kubenswrapper[4676]: I0124 00:31:41.046690 4676 scope.go:117] "RemoveContainer" containerID="3555dab39dcc24cb3d738eb810520147594f4722ea764e376bb893954691a29d" Jan 24 00:31:42 crc kubenswrapper[4676]: I0124 00:31:42.256526 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:31:42 crc kubenswrapper[4676]: E0124 00:31:42.257183 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.057695 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d058-account-create-update-pjzm7"] Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.076741 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-4766-account-create-update-bbg95"] Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.088601 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-b4lc5"] Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.095733 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d058-account-create-update-pjzm7"] Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.102727 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-ba07-account-create-update-hw2sx"] Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.110583 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-b4lc5"] Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.117060 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-ba07-account-create-update-hw2sx"] Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.123140 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-4766-account-create-update-bbg95"] Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.130157 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-mtdlz"] Jan 24 00:31:45 crc kubenswrapper[4676]: I0124 00:31:45.136598 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-mtdlz"] Jan 24 00:31:46 crc kubenswrapper[4676]: I0124 00:31:46.267478 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d4fb3ef-0cde-499c-8018-14cc96b495f5" path="/var/lib/kubelet/pods/2d4fb3ef-0cde-499c-8018-14cc96b495f5/volumes" Jan 24 00:31:46 crc kubenswrapper[4676]: I0124 00:31:46.269578 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64a07672-60fb-4935-83e2-99f39e15427f" path="/var/lib/kubelet/pods/64a07672-60fb-4935-83e2-99f39e15427f/volumes" Jan 24 00:31:46 crc kubenswrapper[4676]: I0124 00:31:46.270437 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64efcf33-9ffe-402d-b0b0-9cf53c7a5495" path="/var/lib/kubelet/pods/64efcf33-9ffe-402d-b0b0-9cf53c7a5495/volumes" Jan 24 00:31:46 crc kubenswrapper[4676]: I0124 00:31:46.271236 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ace7871-3944-4c72-980f-c9d5e7d65c71" path="/var/lib/kubelet/pods/9ace7871-3944-4c72-980f-c9d5e7d65c71/volumes" Jan 24 00:31:46 crc kubenswrapper[4676]: I0124 00:31:46.272485 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a61f15c0-2fa1-4d69-bb80-471d6e4e7e09" path="/var/lib/kubelet/pods/a61f15c0-2fa1-4d69-bb80-471d6e4e7e09/volumes" Jan 24 00:31:50 crc kubenswrapper[4676]: I0124 00:31:50.056868 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wg6tc"] Jan 24 00:31:50 crc kubenswrapper[4676]: I0124 00:31:50.080548 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wg6tc"] Jan 24 00:31:50 crc kubenswrapper[4676]: I0124 00:31:50.268933 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30027bc7-2637-4abb-9568-1190cf80a3a5" path="/var/lib/kubelet/pods/30027bc7-2637-4abb-9568-1190cf80a3a5/volumes" Jan 24 00:31:53 crc kubenswrapper[4676]: I0124 00:31:53.256853 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:31:53 crc kubenswrapper[4676]: E0124 00:31:53.257590 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:32:07 crc kubenswrapper[4676]: I0124 00:32:07.255361 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:32:07 crc kubenswrapper[4676]: E0124 00:32:07.256137 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:32:22 crc kubenswrapper[4676]: I0124 00:32:22.255811 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:32:22 crc kubenswrapper[4676]: E0124 00:32:22.256827 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:32:25 crc kubenswrapper[4676]: I0124 00:32:25.054022 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-bpbdz"] Jan 24 00:32:25 crc kubenswrapper[4676]: I0124 00:32:25.068150 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-bpbdz"] Jan 24 00:32:26 crc kubenswrapper[4676]: I0124 00:32:26.266113 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01c23d24-6fd8-4660-a0f7-815fdf508f5b" path="/var/lib/kubelet/pods/01c23d24-6fd8-4660-a0f7-815fdf508f5b/volumes" Jan 24 00:32:30 crc kubenswrapper[4676]: I0124 00:32:30.029749 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-4kdvp"] Jan 24 00:32:30 crc kubenswrapper[4676]: I0124 00:32:30.036284 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-4kdvp"] Jan 24 00:32:30 crc kubenswrapper[4676]: I0124 00:32:30.270528 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c17fc5e6-983e-4678-b22c-c68686271163" path="/var/lib/kubelet/pods/c17fc5e6-983e-4678-b22c-c68686271163/volumes" Jan 24 00:32:35 crc kubenswrapper[4676]: I0124 00:32:35.051874 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-l49lm"] Jan 24 00:32:35 crc kubenswrapper[4676]: I0124 00:32:35.062837 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-l49lm"] Jan 24 00:32:36 crc kubenswrapper[4676]: I0124 00:32:36.024767 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-g5dnq"] Jan 24 00:32:36 crc kubenswrapper[4676]: I0124 00:32:36.035145 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-g5dnq"] Jan 24 00:32:36 crc kubenswrapper[4676]: I0124 00:32:36.274831 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04db8ba1-c0de-4985-8a48-ead625786472" path="/var/lib/kubelet/pods/04db8ba1-c0de-4985-8a48-ead625786472/volumes" Jan 24 00:32:36 crc kubenswrapper[4676]: I0124 00:32:36.275754 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7327fe7-179a-4492-8823-94b1067c17d4" path="/var/lib/kubelet/pods/a7327fe7-179a-4492-8823-94b1067c17d4/volumes" Jan 24 00:32:37 crc kubenswrapper[4676]: I0124 00:32:37.257149 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:32:37 crc kubenswrapper[4676]: E0124 00:32:37.257801 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.234762 4676 scope.go:117] "RemoveContainer" containerID="531b26a23fe05a6b22c6e35ea5ff32eb5bb18c34495f3ce046a990e3e2b684b5" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.258281 4676 scope.go:117] "RemoveContainer" containerID="192824bb8ad730af6c36357704c067406da5b951731ba1841c964e1a78d10bc2" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.303883 4676 scope.go:117] "RemoveContainer" containerID="47e15b8611d25a16d27aae65f29f85b602c09a46ed5cfeeea3914ac200a1d0a7" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.349796 4676 scope.go:117] "RemoveContainer" containerID="acb8fc5743669070f34cea9944d1cdf719e2212dfd00759a3d455ff74f1aa3b7" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.381311 4676 scope.go:117] "RemoveContainer" containerID="c87e9f17f624e372e522a3229d29135c0ab3c4953ce39a4a0c79e4002fe69658" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.453569 4676 scope.go:117] "RemoveContainer" containerID="9f879edfb9e19e2ba5d1f7acc73cb74d671c4cd334c91721768ab3acc3bba0e9" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.478415 4676 scope.go:117] "RemoveContainer" containerID="b062d260d4a7b754ef8f1768435c8c0945467373bcd6d26d2a060794a30067d3" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.503483 4676 scope.go:117] "RemoveContainer" containerID="2562adc9221a7b111bd58848de15d9abcd709528ca168c846292e16b34c5154e" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.528935 4676 scope.go:117] "RemoveContainer" containerID="c0f6b7c08c73c5cd3ba2389c70cde6261bfdc95261dc1f7fcb14c46b9908056d" Jan 24 00:32:41 crc kubenswrapper[4676]: I0124 00:32:41.553790 4676 scope.go:117] "RemoveContainer" containerID="bfee0ea90c79ca47b4f4a9473a613bbee236021e6a49db23a4490b8cde0e9a39" Jan 24 00:32:51 crc kubenswrapper[4676]: I0124 00:32:51.063979 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7klnv"] Jan 24 00:32:51 crc kubenswrapper[4676]: I0124 00:32:51.076462 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7klnv"] Jan 24 00:32:51 crc kubenswrapper[4676]: I0124 00:32:51.256712 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:32:51 crc kubenswrapper[4676]: E0124 00:32:51.256912 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:32:52 crc kubenswrapper[4676]: I0124 00:32:52.265450 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1163cc6-7ce1-4f2b-9d4a-f3e215177842" path="/var/lib/kubelet/pods/d1163cc6-7ce1-4f2b-9d4a-f3e215177842/volumes" Jan 24 00:33:05 crc kubenswrapper[4676]: I0124 00:33:05.926221 4676 generic.go:334] "Generic (PLEG): container finished" podID="7e20d58a-5f01-4c23-9ab0-650d3ae76844" containerID="a00ce8325a13556bd7fc250572ccaf4efe71b1b28dfcc05e56a4b357eae4bb8d" exitCode=0 Jan 24 00:33:05 crc kubenswrapper[4676]: I0124 00:33:05.926319 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" event={"ID":"7e20d58a-5f01-4c23-9ab0-650d3ae76844","Type":"ContainerDied","Data":"a00ce8325a13556bd7fc250572ccaf4efe71b1b28dfcc05e56a4b357eae4bb8d"} Jan 24 00:33:06 crc kubenswrapper[4676]: I0124 00:33:06.267697 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:33:06 crc kubenswrapper[4676]: E0124 00:33:06.267942 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.369570 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.488572 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-ssh-key-openstack-edpm-ipam\") pod \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.488860 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-inventory\") pod \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.488973 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22ck7\" (UniqueName: \"kubernetes.io/projected/7e20d58a-5f01-4c23-9ab0-650d3ae76844-kube-api-access-22ck7\") pod \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\" (UID: \"7e20d58a-5f01-4c23-9ab0-650d3ae76844\") " Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.507985 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e20d58a-5f01-4c23-9ab0-650d3ae76844-kube-api-access-22ck7" (OuterVolumeSpecName: "kube-api-access-22ck7") pod "7e20d58a-5f01-4c23-9ab0-650d3ae76844" (UID: "7e20d58a-5f01-4c23-9ab0-650d3ae76844"). InnerVolumeSpecName "kube-api-access-22ck7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.525619 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-inventory" (OuterVolumeSpecName: "inventory") pod "7e20d58a-5f01-4c23-9ab0-650d3ae76844" (UID: "7e20d58a-5f01-4c23-9ab0-650d3ae76844"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.527881 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7e20d58a-5f01-4c23-9ab0-650d3ae76844" (UID: "7e20d58a-5f01-4c23-9ab0-650d3ae76844"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.591646 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.591688 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22ck7\" (UniqueName: \"kubernetes.io/projected/7e20d58a-5f01-4c23-9ab0-650d3ae76844-kube-api-access-22ck7\") on node \"crc\" DevicePath \"\"" Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.591703 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7e20d58a-5f01-4c23-9ab0-650d3ae76844-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.951355 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" event={"ID":"7e20d58a-5f01-4c23-9ab0-650d3ae76844","Type":"ContainerDied","Data":"d5ce814e830d10bf24b9a1a0312e8af78b1ad869f8900e1ffd1c55e849d5a89e"} Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.951569 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5ce814e830d10bf24b9a1a0312e8af78b1ad869f8900e1ffd1c55e849d5a89e" Jan 24 00:33:07 crc kubenswrapper[4676]: I0124 00:33:07.951485 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-skh8k" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.086897 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg"] Jan 24 00:33:08 crc kubenswrapper[4676]: E0124 00:33:08.087328 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e20d58a-5f01-4c23-9ab0-650d3ae76844" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.087345 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e20d58a-5f01-4c23-9ab0-650d3ae76844" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.087564 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e20d58a-5f01-4c23-9ab0-650d3ae76844" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.088167 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.091725 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.091859 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.092050 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.092159 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.103663 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg"] Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.205694 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4nq8\" (UniqueName: \"kubernetes.io/projected/7f552002-93ef-485f-9227-a94733534466-kube-api-access-d4nq8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.205894 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.206174 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.307524 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.307595 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4nq8\" (UniqueName: \"kubernetes.io/projected/7f552002-93ef-485f-9227-a94733534466-kube-api-access-d4nq8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.307697 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.311367 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.315225 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.322338 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4nq8\" (UniqueName: \"kubernetes.io/projected/7f552002-93ef-485f-9227-a94733534466-kube-api-access-d4nq8\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:08 crc kubenswrapper[4676]: I0124 00:33:08.432438 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:33:09 crc kubenswrapper[4676]: I0124 00:33:09.055263 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg"] Jan 24 00:33:10 crc kubenswrapper[4676]: I0124 00:33:10.010426 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" event={"ID":"7f552002-93ef-485f-9227-a94733534466","Type":"ContainerStarted","Data":"ed3d716a27de10e6db253fd9bf5e4076a67003feba735058d5c8105d10b30d15"} Jan 24 00:33:11 crc kubenswrapper[4676]: I0124 00:33:11.021640 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" event={"ID":"7f552002-93ef-485f-9227-a94733534466","Type":"ContainerStarted","Data":"c5211d8770447c14431877d5448ca4697609d313cfeb7328a9986f02813d587d"} Jan 24 00:33:11 crc kubenswrapper[4676]: I0124 00:33:11.055001 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" podStartSLOduration=2.060294489 podStartE2EDuration="3.05389409s" podCreationTimestamp="2026-01-24 00:33:08 +0000 UTC" firstStartedPulling="2026-01-24 00:33:09.039054811 +0000 UTC m=+1773.069025822" lastFinishedPulling="2026-01-24 00:33:10.032654412 +0000 UTC m=+1774.062625423" observedRunningTime="2026-01-24 00:33:11.038854338 +0000 UTC m=+1775.068825379" watchObservedRunningTime="2026-01-24 00:33:11.05389409 +0000 UTC m=+1775.083865131" Jan 24 00:33:20 crc kubenswrapper[4676]: I0124 00:33:20.256538 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:33:20 crc kubenswrapper[4676]: E0124 00:33:20.257520 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:33:29 crc kubenswrapper[4676]: I0124 00:33:29.048218 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-95bf-account-create-update-7cvdb"] Jan 24 00:33:29 crc kubenswrapper[4676]: I0124 00:33:29.060613 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-95bf-account-create-update-7cvdb"] Jan 24 00:33:30 crc kubenswrapper[4676]: I0124 00:33:30.048033 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-a963-account-create-update-jx4g6"] Jan 24 00:33:30 crc kubenswrapper[4676]: I0124 00:33:30.066012 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-a963-account-create-update-jx4g6"] Jan 24 00:33:30 crc kubenswrapper[4676]: I0124 00:33:30.283483 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a5f8ec9-ee62-4078-846c-291a47631ffb" path="/var/lib/kubelet/pods/2a5f8ec9-ee62-4078-846c-291a47631ffb/volumes" Jan 24 00:33:30 crc kubenswrapper[4676]: I0124 00:33:30.288999 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d47faf0-8eaa-474f-8cde-aed6c10f4a05" path="/var/lib/kubelet/pods/7d47faf0-8eaa-474f-8cde-aed6c10f4a05/volumes" Jan 24 00:33:31 crc kubenswrapper[4676]: I0124 00:33:31.046871 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-vtxzm"] Jan 24 00:33:31 crc kubenswrapper[4676]: I0124 00:33:31.066995 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-bpr95"] Jan 24 00:33:31 crc kubenswrapper[4676]: I0124 00:33:31.076846 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-b7wzs"] Jan 24 00:33:31 crc kubenswrapper[4676]: I0124 00:33:31.086835 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-bpr95"] Jan 24 00:33:31 crc kubenswrapper[4676]: I0124 00:33:31.095179 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-vtxzm"] Jan 24 00:33:31 crc kubenswrapper[4676]: I0124 00:33:31.103697 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-b7wzs"] Jan 24 00:33:31 crc kubenswrapper[4676]: I0124 00:33:31.111161 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5f9c-account-create-update-9tmf6"] Jan 24 00:33:31 crc kubenswrapper[4676]: I0124 00:33:31.117803 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5f9c-account-create-update-9tmf6"] Jan 24 00:33:31 crc kubenswrapper[4676]: I0124 00:33:31.274050 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:33:31 crc kubenswrapper[4676]: E0124 00:33:31.274781 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:33:32 crc kubenswrapper[4676]: I0124 00:33:32.275524 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="792cc535-5ed4-4e06-a19c-31ba34c7dfc7" path="/var/lib/kubelet/pods/792cc535-5ed4-4e06-a19c-31ba34c7dfc7/volumes" Jan 24 00:33:32 crc kubenswrapper[4676]: I0124 00:33:32.276861 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cd4cfd3-72ab-45d6-8683-85acb6cadf66" path="/var/lib/kubelet/pods/9cd4cfd3-72ab-45d6-8683-85acb6cadf66/volumes" Jan 24 00:33:32 crc kubenswrapper[4676]: I0124 00:33:32.278076 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd4662bb-5179-4f57-8571-5198dcf69bdb" path="/var/lib/kubelet/pods/bd4662bb-5179-4f57-8571-5198dcf69bdb/volumes" Jan 24 00:33:32 crc kubenswrapper[4676]: I0124 00:33:32.279233 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee552380-8d96-4a10-a5b8-2deb2e73b15f" path="/var/lib/kubelet/pods/ee552380-8d96-4a10-a5b8-2deb2e73b15f/volumes" Jan 24 00:33:41 crc kubenswrapper[4676]: I0124 00:33:41.752614 4676 scope.go:117] "RemoveContainer" containerID="cc3b8c014c9450ef0997c6ac2396003671e4bb4fb9c8343ce0debdf9d3796b15" Jan 24 00:33:41 crc kubenswrapper[4676]: I0124 00:33:41.781628 4676 scope.go:117] "RemoveContainer" containerID="e7fc18fa4adecb5d9d62a155b3da665c4c438818bc9b1324ec05947c45e02af0" Jan 24 00:33:41 crc kubenswrapper[4676]: I0124 00:33:41.828818 4676 scope.go:117] "RemoveContainer" containerID="3ee83666354e8c207624c3af1d827bd9c61811a7455e005083ff4e2e34ef0d81" Jan 24 00:33:41 crc kubenswrapper[4676]: I0124 00:33:41.876132 4676 scope.go:117] "RemoveContainer" containerID="e4299d6b00616548489e178e72141c6cc9961df799e192961e6eb82793cfee65" Jan 24 00:33:41 crc kubenswrapper[4676]: I0124 00:33:41.922548 4676 scope.go:117] "RemoveContainer" containerID="27297ea6827aaabd3e76dbca0980a960477f82e37ca6b6bb47dba9c165652738" Jan 24 00:33:41 crc kubenswrapper[4676]: I0124 00:33:41.967286 4676 scope.go:117] "RemoveContainer" containerID="848f0e8f0de2d0ecec1629bf16e2fb50f2be26fc04335b7c112ae212cbf9035a" Jan 24 00:33:42 crc kubenswrapper[4676]: I0124 00:33:42.017169 4676 scope.go:117] "RemoveContainer" containerID="0b76bccfcacd7fb1c879fa9b8c0e01b681d270afe951e54be9117c071c170135" Jan 24 00:33:45 crc kubenswrapper[4676]: I0124 00:33:45.256474 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:33:45 crc kubenswrapper[4676]: E0124 00:33:45.257268 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:33:58 crc kubenswrapper[4676]: I0124 00:33:58.257025 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:33:58 crc kubenswrapper[4676]: E0124 00:33:58.257829 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:34:05 crc kubenswrapper[4676]: I0124 00:34:05.057888 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ptgmr"] Jan 24 00:34:05 crc kubenswrapper[4676]: I0124 00:34:05.070704 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ptgmr"] Jan 24 00:34:06 crc kubenswrapper[4676]: I0124 00:34:06.269999 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff" path="/var/lib/kubelet/pods/9d3caecf-1bb3-49da-ab6a-76d7cf0e08ff/volumes" Jan 24 00:34:13 crc kubenswrapper[4676]: I0124 00:34:13.256572 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:34:13 crc kubenswrapper[4676]: E0124 00:34:13.257321 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:34:26 crc kubenswrapper[4676]: I0124 00:34:26.262954 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:34:26 crc kubenswrapper[4676]: E0124 00:34:26.264957 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:34:30 crc kubenswrapper[4676]: I0124 00:34:30.045409 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-2cwbv"] Jan 24 00:34:30 crc kubenswrapper[4676]: I0124 00:34:30.052455 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fpw6c"] Jan 24 00:34:30 crc kubenswrapper[4676]: I0124 00:34:30.057485 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-2cwbv"] Jan 24 00:34:30 crc kubenswrapper[4676]: I0124 00:34:30.065137 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-fpw6c"] Jan 24 00:34:30 crc kubenswrapper[4676]: I0124 00:34:30.275285 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ee94537-f601-4765-95cb-d56518fb7fc6" path="/var/lib/kubelet/pods/8ee94537-f601-4765-95cb-d56518fb7fc6/volumes" Jan 24 00:34:30 crc kubenswrapper[4676]: I0124 00:34:30.276126 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1b05e80-51fb-476f-a2b4-bf5a290ea5ae" path="/var/lib/kubelet/pods/a1b05e80-51fb-476f-a2b4-bf5a290ea5ae/volumes" Jan 24 00:34:36 crc kubenswrapper[4676]: I0124 00:34:36.850047 4676 generic.go:334] "Generic (PLEG): container finished" podID="7f552002-93ef-485f-9227-a94733534466" containerID="c5211d8770447c14431877d5448ca4697609d313cfeb7328a9986f02813d587d" exitCode=0 Jan 24 00:34:36 crc kubenswrapper[4676]: I0124 00:34:36.850230 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" event={"ID":"7f552002-93ef-485f-9227-a94733534466","Type":"ContainerDied","Data":"c5211d8770447c14431877d5448ca4697609d313cfeb7328a9986f02813d587d"} Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.255899 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:34:38 crc kubenswrapper[4676]: E0124 00:34:38.256656 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.283640 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.397612 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-ssh-key-openstack-edpm-ipam\") pod \"7f552002-93ef-485f-9227-a94733534466\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.398055 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-inventory\") pod \"7f552002-93ef-485f-9227-a94733534466\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.398131 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4nq8\" (UniqueName: \"kubernetes.io/projected/7f552002-93ef-485f-9227-a94733534466-kube-api-access-d4nq8\") pod \"7f552002-93ef-485f-9227-a94733534466\" (UID: \"7f552002-93ef-485f-9227-a94733534466\") " Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.406885 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f552002-93ef-485f-9227-a94733534466-kube-api-access-d4nq8" (OuterVolumeSpecName: "kube-api-access-d4nq8") pod "7f552002-93ef-485f-9227-a94733534466" (UID: "7f552002-93ef-485f-9227-a94733534466"). InnerVolumeSpecName "kube-api-access-d4nq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.422484 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7f552002-93ef-485f-9227-a94733534466" (UID: "7f552002-93ef-485f-9227-a94733534466"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.427844 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-inventory" (OuterVolumeSpecName: "inventory") pod "7f552002-93ef-485f-9227-a94733534466" (UID: "7f552002-93ef-485f-9227-a94733534466"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.500110 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.500140 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4nq8\" (UniqueName: \"kubernetes.io/projected/7f552002-93ef-485f-9227-a94733534466-kube-api-access-d4nq8\") on node \"crc\" DevicePath \"\"" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.500152 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7f552002-93ef-485f-9227-a94733534466-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.890523 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" event={"ID":"7f552002-93ef-485f-9227-a94733534466","Type":"ContainerDied","Data":"ed3d716a27de10e6db253fd9bf5e4076a67003feba735058d5c8105d10b30d15"} Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.890568 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed3d716a27de10e6db253fd9bf5e4076a67003feba735058d5c8105d10b30d15" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.890602 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.992521 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd"] Jan 24 00:34:38 crc kubenswrapper[4676]: E0124 00:34:38.992934 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f552002-93ef-485f-9227-a94733534466" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.992949 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f552002-93ef-485f-9227-a94733534466" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.994539 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f552002-93ef-485f-9227-a94733534466" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Jan 24 00:34:38 crc kubenswrapper[4676]: I0124 00:34:38.995308 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.003243 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.003399 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.003525 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.003627 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.012737 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd"] Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.110897 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.111217 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4r55\" (UniqueName: \"kubernetes.io/projected/83640815-cc06-4abe-a06f-20a1f8798609-kube-api-access-f4r55\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.111278 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.212606 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.212671 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4r55\" (UniqueName: \"kubernetes.io/projected/83640815-cc06-4abe-a06f-20a1f8798609-kube-api-access-f4r55\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.212724 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.217519 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.227993 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.238119 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4r55\" (UniqueName: \"kubernetes.io/projected/83640815-cc06-4abe-a06f-20a1f8798609-kube-api-access-f4r55\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.319202 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:39 crc kubenswrapper[4676]: I0124 00:34:39.920457 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd"] Jan 24 00:34:40 crc kubenswrapper[4676]: I0124 00:34:40.909099 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" event={"ID":"83640815-cc06-4abe-a06f-20a1f8798609","Type":"ContainerStarted","Data":"053dee6005bb4898c9263694d9cc519e8c586b5f9e5a44d6b717d52a218b5d54"} Jan 24 00:34:40 crc kubenswrapper[4676]: I0124 00:34:40.909352 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" event={"ID":"83640815-cc06-4abe-a06f-20a1f8798609","Type":"ContainerStarted","Data":"079bb0127e01a451f6f4f9fc2bbb1280917d9b125ded43402f6e87d34026dee7"} Jan 24 00:34:40 crc kubenswrapper[4676]: I0124 00:34:40.940248 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" podStartSLOduration=2.368930726 podStartE2EDuration="2.940229528s" podCreationTimestamp="2026-01-24 00:34:38 +0000 UTC" firstStartedPulling="2026-01-24 00:34:39.929708129 +0000 UTC m=+1863.959679140" lastFinishedPulling="2026-01-24 00:34:40.501006941 +0000 UTC m=+1864.530977942" observedRunningTime="2026-01-24 00:34:40.93672745 +0000 UTC m=+1864.966698451" watchObservedRunningTime="2026-01-24 00:34:40.940229528 +0000 UTC m=+1864.970200529" Jan 24 00:34:42 crc kubenswrapper[4676]: I0124 00:34:42.153325 4676 scope.go:117] "RemoveContainer" containerID="78da04e8326c020891ea05f676639495efe89b95f4c1821c5af31b8929259b96" Jan 24 00:34:42 crc kubenswrapper[4676]: I0124 00:34:42.183496 4676 scope.go:117] "RemoveContainer" containerID="1e6d788c12343033e8113f7b2adf05ab1c0f4bf8cb67dce88248d068873baa87" Jan 24 00:34:42 crc kubenswrapper[4676]: I0124 00:34:42.230630 4676 scope.go:117] "RemoveContainer" containerID="ba2f94acd9876a58363d3def84609810640e18145307211383bf62958eaadb26" Jan 24 00:34:46 crc kubenswrapper[4676]: I0124 00:34:46.969739 4676 generic.go:334] "Generic (PLEG): container finished" podID="83640815-cc06-4abe-a06f-20a1f8798609" containerID="053dee6005bb4898c9263694d9cc519e8c586b5f9e5a44d6b717d52a218b5d54" exitCode=0 Jan 24 00:34:46 crc kubenswrapper[4676]: I0124 00:34:46.969816 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" event={"ID":"83640815-cc06-4abe-a06f-20a1f8798609","Type":"ContainerDied","Data":"053dee6005bb4898c9263694d9cc519e8c586b5f9e5a44d6b717d52a218b5d54"} Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.381161 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.504049 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-ssh-key-openstack-edpm-ipam\") pod \"83640815-cc06-4abe-a06f-20a1f8798609\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.504144 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-inventory\") pod \"83640815-cc06-4abe-a06f-20a1f8798609\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.504265 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4r55\" (UniqueName: \"kubernetes.io/projected/83640815-cc06-4abe-a06f-20a1f8798609-kube-api-access-f4r55\") pod \"83640815-cc06-4abe-a06f-20a1f8798609\" (UID: \"83640815-cc06-4abe-a06f-20a1f8798609\") " Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.516419 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83640815-cc06-4abe-a06f-20a1f8798609-kube-api-access-f4r55" (OuterVolumeSpecName: "kube-api-access-f4r55") pod "83640815-cc06-4abe-a06f-20a1f8798609" (UID: "83640815-cc06-4abe-a06f-20a1f8798609"). InnerVolumeSpecName "kube-api-access-f4r55". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.534409 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-inventory" (OuterVolumeSpecName: "inventory") pod "83640815-cc06-4abe-a06f-20a1f8798609" (UID: "83640815-cc06-4abe-a06f-20a1f8798609"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.539484 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "83640815-cc06-4abe-a06f-20a1f8798609" (UID: "83640815-cc06-4abe-a06f-20a1f8798609"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.606488 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.606520 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83640815-cc06-4abe-a06f-20a1f8798609-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:34:48 crc kubenswrapper[4676]: I0124 00:34:48.606531 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4r55\" (UniqueName: \"kubernetes.io/projected/83640815-cc06-4abe-a06f-20a1f8798609-kube-api-access-f4r55\") on node \"crc\" DevicePath \"\"" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.033902 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" event={"ID":"83640815-cc06-4abe-a06f-20a1f8798609","Type":"ContainerDied","Data":"079bb0127e01a451f6f4f9fc2bbb1280917d9b125ded43402f6e87d34026dee7"} Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.033949 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="079bb0127e01a451f6f4f9fc2bbb1280917d9b125ded43402f6e87d34026dee7" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.034078 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.108026 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c"] Jan 24 00:34:49 crc kubenswrapper[4676]: E0124 00:34:49.108404 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83640815-cc06-4abe-a06f-20a1f8798609" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.108422 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="83640815-cc06-4abe-a06f-20a1f8798609" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.108730 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="83640815-cc06-4abe-a06f-20a1f8798609" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.109644 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c"] Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.109740 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.116242 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.116485 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.116573 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.116612 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.219326 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sz24c\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.219395 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrjdn\" (UniqueName: \"kubernetes.io/projected/15fbef8d-e606-4b72-a994-e71d03e8fec8-kube-api-access-qrjdn\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sz24c\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.219494 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sz24c\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.321441 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sz24c\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.322183 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sz24c\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.322281 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrjdn\" (UniqueName: \"kubernetes.io/projected/15fbef8d-e606-4b72-a994-e71d03e8fec8-kube-api-access-qrjdn\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sz24c\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.329327 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sz24c\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.330935 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sz24c\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.339066 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrjdn\" (UniqueName: \"kubernetes.io/projected/15fbef8d-e606-4b72-a994-e71d03e8fec8-kube-api-access-qrjdn\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sz24c\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:49 crc kubenswrapper[4676]: I0124 00:34:49.438839 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:34:50 crc kubenswrapper[4676]: I0124 00:34:50.002837 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c"] Jan 24 00:34:50 crc kubenswrapper[4676]: I0124 00:34:50.045152 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" event={"ID":"15fbef8d-e606-4b72-a994-e71d03e8fec8","Type":"ContainerStarted","Data":"7ba6922eccbf3af44bee7c474d3f94451469a6bfb3124c944eb7ac801e95824b"} Jan 24 00:34:51 crc kubenswrapper[4676]: I0124 00:34:51.056027 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" event={"ID":"15fbef8d-e606-4b72-a994-e71d03e8fec8","Type":"ContainerStarted","Data":"f5244c90e6a886fde151ab47a266d37ba45ab65855478b59a54b16449bd8cac9"} Jan 24 00:34:51 crc kubenswrapper[4676]: I0124 00:34:51.081977 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" podStartSLOduration=1.610067049 podStartE2EDuration="2.081951506s" podCreationTimestamp="2026-01-24 00:34:49 +0000 UTC" firstStartedPulling="2026-01-24 00:34:50.012677117 +0000 UTC m=+1874.042648158" lastFinishedPulling="2026-01-24 00:34:50.484561614 +0000 UTC m=+1874.514532615" observedRunningTime="2026-01-24 00:34:51.07489861 +0000 UTC m=+1875.104869631" watchObservedRunningTime="2026-01-24 00:34:51.081951506 +0000 UTC m=+1875.111922517" Jan 24 00:34:51 crc kubenswrapper[4676]: I0124 00:34:51.256456 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:34:51 crc kubenswrapper[4676]: E0124 00:34:51.256758 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:35:06 crc kubenswrapper[4676]: I0124 00:35:06.270109 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:35:06 crc kubenswrapper[4676]: E0124 00:35:06.270738 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:35:14 crc kubenswrapper[4676]: I0124 00:35:14.041994 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-49926"] Jan 24 00:35:14 crc kubenswrapper[4676]: I0124 00:35:14.051819 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-49926"] Jan 24 00:35:14 crc kubenswrapper[4676]: I0124 00:35:14.265700 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e6d63aa-e36a-4ef7-b50d-255b44e72c20" path="/var/lib/kubelet/pods/2e6d63aa-e36a-4ef7-b50d-255b44e72c20/volumes" Jan 24 00:35:17 crc kubenswrapper[4676]: I0124 00:35:17.255174 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:35:18 crc kubenswrapper[4676]: I0124 00:35:18.309852 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"658098ecc7ebeb43955aee4f3317cbcc452c89ec3e2a5ddc24ed196ea90b98d2"} Jan 24 00:35:37 crc kubenswrapper[4676]: I0124 00:35:37.486252 4676 generic.go:334] "Generic (PLEG): container finished" podID="15fbef8d-e606-4b72-a994-e71d03e8fec8" containerID="f5244c90e6a886fde151ab47a266d37ba45ab65855478b59a54b16449bd8cac9" exitCode=0 Jan 24 00:35:37 crc kubenswrapper[4676]: I0124 00:35:37.486355 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" event={"ID":"15fbef8d-e606-4b72-a994-e71d03e8fec8","Type":"ContainerDied","Data":"f5244c90e6a886fde151ab47a266d37ba45ab65855478b59a54b16449bd8cac9"} Jan 24 00:35:38 crc kubenswrapper[4676]: I0124 00:35:38.898974 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.067116 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrjdn\" (UniqueName: \"kubernetes.io/projected/15fbef8d-e606-4b72-a994-e71d03e8fec8-kube-api-access-qrjdn\") pod \"15fbef8d-e606-4b72-a994-e71d03e8fec8\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.067183 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-ssh-key-openstack-edpm-ipam\") pod \"15fbef8d-e606-4b72-a994-e71d03e8fec8\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.067365 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-inventory\") pod \"15fbef8d-e606-4b72-a994-e71d03e8fec8\" (UID: \"15fbef8d-e606-4b72-a994-e71d03e8fec8\") " Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.072912 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15fbef8d-e606-4b72-a994-e71d03e8fec8-kube-api-access-qrjdn" (OuterVolumeSpecName: "kube-api-access-qrjdn") pod "15fbef8d-e606-4b72-a994-e71d03e8fec8" (UID: "15fbef8d-e606-4b72-a994-e71d03e8fec8"). InnerVolumeSpecName "kube-api-access-qrjdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.100634 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-inventory" (OuterVolumeSpecName: "inventory") pod "15fbef8d-e606-4b72-a994-e71d03e8fec8" (UID: "15fbef8d-e606-4b72-a994-e71d03e8fec8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.108626 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "15fbef8d-e606-4b72-a994-e71d03e8fec8" (UID: "15fbef8d-e606-4b72-a994-e71d03e8fec8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.170161 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrjdn\" (UniqueName: \"kubernetes.io/projected/15fbef8d-e606-4b72-a994-e71d03e8fec8-kube-api-access-qrjdn\") on node \"crc\" DevicePath \"\"" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.170231 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.170246 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/15fbef8d-e606-4b72-a994-e71d03e8fec8-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.509102 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" event={"ID":"15fbef8d-e606-4b72-a994-e71d03e8fec8","Type":"ContainerDied","Data":"7ba6922eccbf3af44bee7c474d3f94451469a6bfb3124c944eb7ac801e95824b"} Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.509158 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ba6922eccbf3af44bee7c474d3f94451469a6bfb3124c944eb7ac801e95824b" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.509182 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sz24c" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.690089 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7"] Jan 24 00:35:39 crc kubenswrapper[4676]: E0124 00:35:39.690782 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15fbef8d-e606-4b72-a994-e71d03e8fec8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.690816 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="15fbef8d-e606-4b72-a994-e71d03e8fec8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.691250 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="15fbef8d-e606-4b72-a994-e71d03e8fec8" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.692102 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.695273 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.698274 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.702452 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.705302 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.724077 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7"] Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.883802 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.883905 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.883934 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x6b4\" (UniqueName: \"kubernetes.io/projected/f6bc5ee4-f730-4e1e-9684-b643daed2519-kube-api-access-9x6b4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.986351 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.986473 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.986511 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x6b4\" (UniqueName: \"kubernetes.io/projected/f6bc5ee4-f730-4e1e-9684-b643daed2519-kube-api-access-9x6b4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.994229 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:39 crc kubenswrapper[4676]: I0124 00:35:39.995962 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:40 crc kubenswrapper[4676]: I0124 00:35:40.004290 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x6b4\" (UniqueName: \"kubernetes.io/projected/f6bc5ee4-f730-4e1e-9684-b643daed2519-kube-api-access-9x6b4\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:40 crc kubenswrapper[4676]: I0124 00:35:40.018267 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:35:40 crc kubenswrapper[4676]: I0124 00:35:40.611853 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7"] Jan 24 00:35:40 crc kubenswrapper[4676]: I0124 00:35:40.619846 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:35:41 crc kubenswrapper[4676]: I0124 00:35:41.525573 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" event={"ID":"f6bc5ee4-f730-4e1e-9684-b643daed2519","Type":"ContainerStarted","Data":"4665658c6d6e687af3b155f8c6276554ef2994904fa73e918324ef0a895dab86"} Jan 24 00:35:41 crc kubenswrapper[4676]: I0124 00:35:41.525853 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" event={"ID":"f6bc5ee4-f730-4e1e-9684-b643daed2519","Type":"ContainerStarted","Data":"9f8fc974d6952d15547a94a1cd02cba6b82b321eee0cdb4aa97b6835f688c7ba"} Jan 24 00:35:42 crc kubenswrapper[4676]: I0124 00:35:42.367713 4676 scope.go:117] "RemoveContainer" containerID="d43689494700a5ceee020f09f4ca2f2fd150c60e2e3c7e0d6cd30145e29f275e" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.708843 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" podStartSLOduration=19.055910678 podStartE2EDuration="19.708822753s" podCreationTimestamp="2026-01-24 00:35:39 +0000 UTC" firstStartedPulling="2026-01-24 00:35:40.619577457 +0000 UTC m=+1924.649548468" lastFinishedPulling="2026-01-24 00:35:41.272489542 +0000 UTC m=+1925.302460543" observedRunningTime="2026-01-24 00:35:41.546792129 +0000 UTC m=+1925.576763130" watchObservedRunningTime="2026-01-24 00:35:58.708822753 +0000 UTC m=+1942.738793754" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.716989 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pjnlx"] Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.721137 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.731835 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pjnlx"] Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.772552 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txfsm\" (UniqueName: \"kubernetes.io/projected/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-kube-api-access-txfsm\") pod \"community-operators-pjnlx\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.772650 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-utilities\") pod \"community-operators-pjnlx\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.772682 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-catalog-content\") pod \"community-operators-pjnlx\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.873836 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-utilities\") pod \"community-operators-pjnlx\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.874063 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-catalog-content\") pod \"community-operators-pjnlx\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.874241 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txfsm\" (UniqueName: \"kubernetes.io/projected/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-kube-api-access-txfsm\") pod \"community-operators-pjnlx\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.874411 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-utilities\") pod \"community-operators-pjnlx\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.874484 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-catalog-content\") pod \"community-operators-pjnlx\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:58 crc kubenswrapper[4676]: I0124 00:35:58.892126 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txfsm\" (UniqueName: \"kubernetes.io/projected/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-kube-api-access-txfsm\") pod \"community-operators-pjnlx\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:59 crc kubenswrapper[4676]: I0124 00:35:59.052741 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:35:59 crc kubenswrapper[4676]: I0124 00:35:59.352088 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pjnlx"] Jan 24 00:35:59 crc kubenswrapper[4676]: I0124 00:35:59.670095 4676 generic.go:334] "Generic (PLEG): container finished" podID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerID="3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66" exitCode=0 Jan 24 00:35:59 crc kubenswrapper[4676]: I0124 00:35:59.670143 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjnlx" event={"ID":"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0","Type":"ContainerDied","Data":"3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66"} Jan 24 00:35:59 crc kubenswrapper[4676]: I0124 00:35:59.670356 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjnlx" event={"ID":"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0","Type":"ContainerStarted","Data":"d0751a2e4fafb0354605e902bc985b1ecae83ff901be37db738f0802bd1d6271"} Jan 24 00:36:01 crc kubenswrapper[4676]: I0124 00:36:01.694231 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjnlx" event={"ID":"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0","Type":"ContainerStarted","Data":"bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b"} Jan 24 00:36:02 crc kubenswrapper[4676]: I0124 00:36:02.704619 4676 generic.go:334] "Generic (PLEG): container finished" podID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerID="bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b" exitCode=0 Jan 24 00:36:02 crc kubenswrapper[4676]: I0124 00:36:02.704689 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjnlx" event={"ID":"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0","Type":"ContainerDied","Data":"bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b"} Jan 24 00:36:03 crc kubenswrapper[4676]: I0124 00:36:03.721963 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjnlx" event={"ID":"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0","Type":"ContainerStarted","Data":"fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b"} Jan 24 00:36:04 crc kubenswrapper[4676]: I0124 00:36:04.752911 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pjnlx" podStartSLOduration=2.922483357 podStartE2EDuration="6.752895155s" podCreationTimestamp="2026-01-24 00:35:58 +0000 UTC" firstStartedPulling="2026-01-24 00:35:59.671875157 +0000 UTC m=+1943.701846158" lastFinishedPulling="2026-01-24 00:36:03.502286955 +0000 UTC m=+1947.532257956" observedRunningTime="2026-01-24 00:36:04.747693015 +0000 UTC m=+1948.777664016" watchObservedRunningTime="2026-01-24 00:36:04.752895155 +0000 UTC m=+1948.782866156" Jan 24 00:36:09 crc kubenswrapper[4676]: I0124 00:36:09.054065 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:36:09 crc kubenswrapper[4676]: I0124 00:36:09.055445 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:36:09 crc kubenswrapper[4676]: I0124 00:36:09.097814 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:36:09 crc kubenswrapper[4676]: I0124 00:36:09.863108 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:36:09 crc kubenswrapper[4676]: I0124 00:36:09.940065 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pjnlx"] Jan 24 00:36:11 crc kubenswrapper[4676]: I0124 00:36:11.797128 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pjnlx" podUID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerName="registry-server" containerID="cri-o://fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b" gracePeriod=2 Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.292115 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.442201 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-utilities\") pod \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.442255 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txfsm\" (UniqueName: \"kubernetes.io/projected/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-kube-api-access-txfsm\") pod \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.442494 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-catalog-content\") pod \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\" (UID: \"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0\") " Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.443907 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-utilities" (OuterVolumeSpecName: "utilities") pod "34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" (UID: "34a715a4-3f3f-49f7-baa8-14d0ad60b0d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.452562 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-kube-api-access-txfsm" (OuterVolumeSpecName: "kube-api-access-txfsm") pod "34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" (UID: "34a715a4-3f3f-49f7-baa8-14d0ad60b0d0"). InnerVolumeSpecName "kube-api-access-txfsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.492706 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" (UID: "34a715a4-3f3f-49f7-baa8-14d0ad60b0d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.544577 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.544611 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txfsm\" (UniqueName: \"kubernetes.io/projected/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-kube-api-access-txfsm\") on node \"crc\" DevicePath \"\"" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.544625 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.810974 4676 generic.go:334] "Generic (PLEG): container finished" podID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerID="fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b" exitCode=0 Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.811016 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjnlx" event={"ID":"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0","Type":"ContainerDied","Data":"fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b"} Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.811043 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjnlx" event={"ID":"34a715a4-3f3f-49f7-baa8-14d0ad60b0d0","Type":"ContainerDied","Data":"d0751a2e4fafb0354605e902bc985b1ecae83ff901be37db738f0802bd1d6271"} Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.811061 4676 scope.go:117] "RemoveContainer" containerID="fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.811102 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjnlx" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.849888 4676 scope.go:117] "RemoveContainer" containerID="bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.878084 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pjnlx"] Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.887807 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pjnlx"] Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.956355 4676 scope.go:117] "RemoveContainer" containerID="3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.985019 4676 scope.go:117] "RemoveContainer" containerID="fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b" Jan 24 00:36:12 crc kubenswrapper[4676]: E0124 00:36:12.985459 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b\": container with ID starting with fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b not found: ID does not exist" containerID="fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.985575 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b"} err="failed to get container status \"fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b\": rpc error: code = NotFound desc = could not find container \"fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b\": container with ID starting with fab27611b7a817d4706a36e474f0e3b4f9ac17514428f2841dc2f9a592827a3b not found: ID does not exist" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.985666 4676 scope.go:117] "RemoveContainer" containerID="bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b" Jan 24 00:36:12 crc kubenswrapper[4676]: E0124 00:36:12.986003 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b\": container with ID starting with bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b not found: ID does not exist" containerID="bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.986093 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b"} err="failed to get container status \"bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b\": rpc error: code = NotFound desc = could not find container \"bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b\": container with ID starting with bf62ab74165fb7568f822f08e9f28b2f740617c77f920d1ed10bb7f56635302b not found: ID does not exist" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.986182 4676 scope.go:117] "RemoveContainer" containerID="3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66" Jan 24 00:36:12 crc kubenswrapper[4676]: E0124 00:36:12.986493 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66\": container with ID starting with 3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66 not found: ID does not exist" containerID="3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66" Jan 24 00:36:12 crc kubenswrapper[4676]: I0124 00:36:12.986570 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66"} err="failed to get container status \"3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66\": rpc error: code = NotFound desc = could not find container \"3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66\": container with ID starting with 3fc10800e2e755474740e8bb060d4da8fa052be7ca48e286e7371b206cd02d66 not found: ID does not exist" Jan 24 00:36:14 crc kubenswrapper[4676]: I0124 00:36:14.271405 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" path="/var/lib/kubelet/pods/34a715a4-3f3f-49f7-baa8-14d0ad60b0d0/volumes" Jan 24 00:36:44 crc kubenswrapper[4676]: I0124 00:36:44.120062 4676 generic.go:334] "Generic (PLEG): container finished" podID="f6bc5ee4-f730-4e1e-9684-b643daed2519" containerID="4665658c6d6e687af3b155f8c6276554ef2994904fa73e918324ef0a895dab86" exitCode=0 Jan 24 00:36:44 crc kubenswrapper[4676]: I0124 00:36:44.120141 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" event={"ID":"f6bc5ee4-f730-4e1e-9684-b643daed2519","Type":"ContainerDied","Data":"4665658c6d6e687af3b155f8c6276554ef2994904fa73e918324ef0a895dab86"} Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.562246 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.617252 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-ssh-key-openstack-edpm-ipam\") pod \"f6bc5ee4-f730-4e1e-9684-b643daed2519\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.617328 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x6b4\" (UniqueName: \"kubernetes.io/projected/f6bc5ee4-f730-4e1e-9684-b643daed2519-kube-api-access-9x6b4\") pod \"f6bc5ee4-f730-4e1e-9684-b643daed2519\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.617517 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-inventory\") pod \"f6bc5ee4-f730-4e1e-9684-b643daed2519\" (UID: \"f6bc5ee4-f730-4e1e-9684-b643daed2519\") " Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.623603 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6bc5ee4-f730-4e1e-9684-b643daed2519-kube-api-access-9x6b4" (OuterVolumeSpecName: "kube-api-access-9x6b4") pod "f6bc5ee4-f730-4e1e-9684-b643daed2519" (UID: "f6bc5ee4-f730-4e1e-9684-b643daed2519"). InnerVolumeSpecName "kube-api-access-9x6b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.651359 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f6bc5ee4-f730-4e1e-9684-b643daed2519" (UID: "f6bc5ee4-f730-4e1e-9684-b643daed2519"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.654218 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-inventory" (OuterVolumeSpecName: "inventory") pod "f6bc5ee4-f730-4e1e-9684-b643daed2519" (UID: "f6bc5ee4-f730-4e1e-9684-b643daed2519"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.720016 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.720050 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x6b4\" (UniqueName: \"kubernetes.io/projected/f6bc5ee4-f730-4e1e-9684-b643daed2519-kube-api-access-9x6b4\") on node \"crc\" DevicePath \"\"" Jan 24 00:36:45 crc kubenswrapper[4676]: I0124 00:36:45.720065 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f6bc5ee4-f730-4e1e-9684-b643daed2519-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.147031 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" event={"ID":"f6bc5ee4-f730-4e1e-9684-b643daed2519","Type":"ContainerDied","Data":"9f8fc974d6952d15547a94a1cd02cba6b82b321eee0cdb4aa97b6835f688c7ba"} Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.147627 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f8fc974d6952d15547a94a1cd02cba6b82b321eee0cdb4aa97b6835f688c7ba" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.147127 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.277154 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8b9rz"] Jan 24 00:36:46 crc kubenswrapper[4676]: E0124 00:36:46.277570 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerName="extract-content" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.277590 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerName="extract-content" Jan 24 00:36:46 crc kubenswrapper[4676]: E0124 00:36:46.277631 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6bc5ee4-f730-4e1e-9684-b643daed2519" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.277642 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6bc5ee4-f730-4e1e-9684-b643daed2519" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:36:46 crc kubenswrapper[4676]: E0124 00:36:46.277690 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerName="extract-utilities" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.277700 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerName="extract-utilities" Jan 24 00:36:46 crc kubenswrapper[4676]: E0124 00:36:46.277716 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerName="registry-server" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.277725 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerName="registry-server" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.277949 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6bc5ee4-f730-4e1e-9684-b643daed2519" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.277982 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a715a4-3f3f-49f7-baa8-14d0ad60b0d0" containerName="registry-server" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.278667 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.288480 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.288513 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.288720 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.289080 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.297685 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8b9rz"] Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.331520 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8b9rz\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.331558 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8b9rz\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.331632 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9z29\" (UniqueName: \"kubernetes.io/projected/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-kube-api-access-d9z29\") pod \"ssh-known-hosts-edpm-deployment-8b9rz\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.432681 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8b9rz\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.432724 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8b9rz\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.432797 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9z29\" (UniqueName: \"kubernetes.io/projected/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-kube-api-access-d9z29\") pod \"ssh-known-hosts-edpm-deployment-8b9rz\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.436453 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8b9rz\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.437601 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8b9rz\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.454174 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9z29\" (UniqueName: \"kubernetes.io/projected/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-kube-api-access-d9z29\") pod \"ssh-known-hosts-edpm-deployment-8b9rz\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:46 crc kubenswrapper[4676]: I0124 00:36:46.631863 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:47 crc kubenswrapper[4676]: I0124 00:36:47.197425 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8b9rz"] Jan 24 00:36:48 crc kubenswrapper[4676]: I0124 00:36:48.168602 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" event={"ID":"d1ed73dc-4392-4d20-a592-4a8c5ba9c104","Type":"ContainerStarted","Data":"e9b15508842a6e3843219b6f379d487b2bda8b91380d62eec7af7fd65f163e7e"} Jan 24 00:36:48 crc kubenswrapper[4676]: I0124 00:36:48.168953 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" event={"ID":"d1ed73dc-4392-4d20-a592-4a8c5ba9c104","Type":"ContainerStarted","Data":"d82373d13c1684a8d66f88055ba4347b10bef0e1909b0ca7e44bf1021f68c9e1"} Jan 24 00:36:48 crc kubenswrapper[4676]: I0124 00:36:48.190543 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" podStartSLOduration=1.726424199 podStartE2EDuration="2.190522376s" podCreationTimestamp="2026-01-24 00:36:46 +0000 UTC" firstStartedPulling="2026-01-24 00:36:47.206867567 +0000 UTC m=+1991.236838558" lastFinishedPulling="2026-01-24 00:36:47.670965724 +0000 UTC m=+1991.700936735" observedRunningTime="2026-01-24 00:36:48.185744949 +0000 UTC m=+1992.215716000" watchObservedRunningTime="2026-01-24 00:36:48.190522376 +0000 UTC m=+1992.220493397" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.211032 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-twx92"] Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.218210 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.273542 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-twx92"] Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.310358 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxlrz\" (UniqueName: \"kubernetes.io/projected/a3e85f87-765d-44bb-9187-6e5276434b05-kube-api-access-fxlrz\") pod \"redhat-operators-twx92\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.310649 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-utilities\") pod \"redhat-operators-twx92\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.310739 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-catalog-content\") pod \"redhat-operators-twx92\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.411864 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxlrz\" (UniqueName: \"kubernetes.io/projected/a3e85f87-765d-44bb-9187-6e5276434b05-kube-api-access-fxlrz\") pod \"redhat-operators-twx92\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.412115 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-utilities\") pod \"redhat-operators-twx92\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.412225 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-catalog-content\") pod \"redhat-operators-twx92\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.412614 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-catalog-content\") pod \"redhat-operators-twx92\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.412706 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-utilities\") pod \"redhat-operators-twx92\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.438762 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxlrz\" (UniqueName: \"kubernetes.io/projected/a3e85f87-765d-44bb-9187-6e5276434b05-kube-api-access-fxlrz\") pod \"redhat-operators-twx92\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:54 crc kubenswrapper[4676]: I0124 00:36:54.564448 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:36:55 crc kubenswrapper[4676]: I0124 00:36:55.238625 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-twx92"] Jan 24 00:36:55 crc kubenswrapper[4676]: I0124 00:36:55.253179 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twx92" event={"ID":"a3e85f87-765d-44bb-9187-6e5276434b05","Type":"ContainerStarted","Data":"12f78548df580a50ff02c0e908a128c3171c6cd389e2bfb1762c569bef21f86a"} Jan 24 00:36:56 crc kubenswrapper[4676]: I0124 00:36:56.271689 4676 generic.go:334] "Generic (PLEG): container finished" podID="a3e85f87-765d-44bb-9187-6e5276434b05" containerID="c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109" exitCode=0 Jan 24 00:36:56 crc kubenswrapper[4676]: I0124 00:36:56.272038 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twx92" event={"ID":"a3e85f87-765d-44bb-9187-6e5276434b05","Type":"ContainerDied","Data":"c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109"} Jan 24 00:36:56 crc kubenswrapper[4676]: I0124 00:36:56.275617 4676 generic.go:334] "Generic (PLEG): container finished" podID="d1ed73dc-4392-4d20-a592-4a8c5ba9c104" containerID="e9b15508842a6e3843219b6f379d487b2bda8b91380d62eec7af7fd65f163e7e" exitCode=0 Jan 24 00:36:56 crc kubenswrapper[4676]: I0124 00:36:56.275670 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" event={"ID":"d1ed73dc-4392-4d20-a592-4a8c5ba9c104","Type":"ContainerDied","Data":"e9b15508842a6e3843219b6f379d487b2bda8b91380d62eec7af7fd65f163e7e"} Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.286248 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twx92" event={"ID":"a3e85f87-765d-44bb-9187-6e5276434b05","Type":"ContainerStarted","Data":"7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c"} Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.756670 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.768327 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-inventory-0\") pod \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.768479 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9z29\" (UniqueName: \"kubernetes.io/projected/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-kube-api-access-d9z29\") pod \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.768607 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-ssh-key-openstack-edpm-ipam\") pod \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\" (UID: \"d1ed73dc-4392-4d20-a592-4a8c5ba9c104\") " Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.778565 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-kube-api-access-d9z29" (OuterVolumeSpecName: "kube-api-access-d9z29") pod "d1ed73dc-4392-4d20-a592-4a8c5ba9c104" (UID: "d1ed73dc-4392-4d20-a592-4a8c5ba9c104"). InnerVolumeSpecName "kube-api-access-d9z29". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.820861 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d1ed73dc-4392-4d20-a592-4a8c5ba9c104" (UID: "d1ed73dc-4392-4d20-a592-4a8c5ba9c104"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.822541 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d1ed73dc-4392-4d20-a592-4a8c5ba9c104" (UID: "d1ed73dc-4392-4d20-a592-4a8c5ba9c104"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.870504 4676 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-inventory-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.870535 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9z29\" (UniqueName: \"kubernetes.io/projected/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-kube-api-access-d9z29\") on node \"crc\" DevicePath \"\"" Jan 24 00:36:57 crc kubenswrapper[4676]: I0124 00:36:57.870547 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d1ed73dc-4392-4d20-a592-4a8c5ba9c104-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.299195 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" event={"ID":"d1ed73dc-4392-4d20-a592-4a8c5ba9c104","Type":"ContainerDied","Data":"d82373d13c1684a8d66f88055ba4347b10bef0e1909b0ca7e44bf1021f68c9e1"} Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.299249 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d82373d13c1684a8d66f88055ba4347b10bef0e1909b0ca7e44bf1021f68c9e1" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.299412 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8b9rz" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.422175 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn"] Jan 24 00:36:58 crc kubenswrapper[4676]: E0124 00:36:58.424264 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ed73dc-4392-4d20-a592-4a8c5ba9c104" containerName="ssh-known-hosts-edpm-deployment" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.424358 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ed73dc-4392-4d20-a592-4a8c5ba9c104" containerName="ssh-known-hosts-edpm-deployment" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.424676 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ed73dc-4392-4d20-a592-4a8c5ba9c104" containerName="ssh-known-hosts-edpm-deployment" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.425397 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.428824 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.429097 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.431299 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.431448 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.439055 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn"] Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.583754 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mhrvn\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.583840 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mhrvn\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.583932 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m52h\" (UniqueName: \"kubernetes.io/projected/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-kube-api-access-9m52h\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mhrvn\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.686247 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mhrvn\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.686353 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mhrvn\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.686452 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m52h\" (UniqueName: \"kubernetes.io/projected/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-kube-api-access-9m52h\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mhrvn\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.711176 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mhrvn\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.713433 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mhrvn\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.713542 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m52h\" (UniqueName: \"kubernetes.io/projected/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-kube-api-access-9m52h\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-mhrvn\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:58 crc kubenswrapper[4676]: I0124 00:36:58.747948 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:36:59 crc kubenswrapper[4676]: I0124 00:36:59.331951 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn"] Jan 24 00:37:00 crc kubenswrapper[4676]: I0124 00:37:00.314112 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" event={"ID":"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1","Type":"ContainerStarted","Data":"084b4e0ab44d55e490dea686f2dbac622f75b83ccd35b40989533f2257cebbae"} Jan 24 00:37:01 crc kubenswrapper[4676]: I0124 00:37:01.326785 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" event={"ID":"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1","Type":"ContainerStarted","Data":"656d4d2d5ff068e82b90b5981ff0e141ca2026de96cf3e479d6668c7e1b25ad7"} Jan 24 00:37:01 crc kubenswrapper[4676]: I0124 00:37:01.329721 4676 generic.go:334] "Generic (PLEG): container finished" podID="a3e85f87-765d-44bb-9187-6e5276434b05" containerID="7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c" exitCode=0 Jan 24 00:37:01 crc kubenswrapper[4676]: I0124 00:37:01.329769 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twx92" event={"ID":"a3e85f87-765d-44bb-9187-6e5276434b05","Type":"ContainerDied","Data":"7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c"} Jan 24 00:37:01 crc kubenswrapper[4676]: I0124 00:37:01.364589 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" podStartSLOduration=1.887540129 podStartE2EDuration="3.364570134s" podCreationTimestamp="2026-01-24 00:36:58 +0000 UTC" firstStartedPulling="2026-01-24 00:36:59.334825407 +0000 UTC m=+2003.364796408" lastFinishedPulling="2026-01-24 00:37:00.811855412 +0000 UTC m=+2004.841826413" observedRunningTime="2026-01-24 00:37:01.354952079 +0000 UTC m=+2005.384923100" watchObservedRunningTime="2026-01-24 00:37:01.364570134 +0000 UTC m=+2005.394541135" Jan 24 00:37:02 crc kubenswrapper[4676]: I0124 00:37:02.356760 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twx92" event={"ID":"a3e85f87-765d-44bb-9187-6e5276434b05","Type":"ContainerStarted","Data":"f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b"} Jan 24 00:37:02 crc kubenswrapper[4676]: I0124 00:37:02.384848 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-twx92" podStartSLOduration=2.8756215689999998 podStartE2EDuration="8.384833279s" podCreationTimestamp="2026-01-24 00:36:54 +0000 UTC" firstStartedPulling="2026-01-24 00:36:56.274296332 +0000 UTC m=+2000.304267333" lastFinishedPulling="2026-01-24 00:37:01.783508032 +0000 UTC m=+2005.813479043" observedRunningTime="2026-01-24 00:37:02.377353669 +0000 UTC m=+2006.407324670" watchObservedRunningTime="2026-01-24 00:37:02.384833279 +0000 UTC m=+2006.414804280" Jan 24 00:37:04 crc kubenswrapper[4676]: I0124 00:37:04.565625 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:37:04 crc kubenswrapper[4676]: I0124 00:37:04.566810 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:37:05 crc kubenswrapper[4676]: I0124 00:37:05.609780 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-twx92" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" containerName="registry-server" probeResult="failure" output=< Jan 24 00:37:05 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:37:05 crc kubenswrapper[4676]: > Jan 24 00:37:11 crc kubenswrapper[4676]: I0124 00:37:11.468335 4676 generic.go:334] "Generic (PLEG): container finished" podID="2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1" containerID="656d4d2d5ff068e82b90b5981ff0e141ca2026de96cf3e479d6668c7e1b25ad7" exitCode=0 Jan 24 00:37:11 crc kubenswrapper[4676]: I0124 00:37:11.468491 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" event={"ID":"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1","Type":"ContainerDied","Data":"656d4d2d5ff068e82b90b5981ff0e141ca2026de96cf3e479d6668c7e1b25ad7"} Jan 24 00:37:12 crc kubenswrapper[4676]: I0124 00:37:12.967808 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.030846 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9m52h\" (UniqueName: \"kubernetes.io/projected/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-kube-api-access-9m52h\") pod \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.030938 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-ssh-key-openstack-edpm-ipam\") pod \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.031137 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-inventory\") pod \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\" (UID: \"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1\") " Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.037561 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-kube-api-access-9m52h" (OuterVolumeSpecName: "kube-api-access-9m52h") pod "2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1" (UID: "2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1"). InnerVolumeSpecName "kube-api-access-9m52h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.057288 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1" (UID: "2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.060701 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-inventory" (OuterVolumeSpecName: "inventory") pod "2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1" (UID: "2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.133659 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.133691 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9m52h\" (UniqueName: \"kubernetes.io/projected/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-kube-api-access-9m52h\") on node \"crc\" DevicePath \"\"" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.133702 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.492967 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" event={"ID":"2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1","Type":"ContainerDied","Data":"084b4e0ab44d55e490dea686f2dbac622f75b83ccd35b40989533f2257cebbae"} Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.493028 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="084b4e0ab44d55e490dea686f2dbac622f75b83ccd35b40989533f2257cebbae" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.493042 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-mhrvn" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.581485 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht"] Jan 24 00:37:13 crc kubenswrapper[4676]: E0124 00:37:13.582076 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.582103 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.582479 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.583527 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.586284 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.586496 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.586795 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.587900 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.596797 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht"] Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.643785 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-448ht\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.643837 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-448ht\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.644314 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z45x6\" (UniqueName: \"kubernetes.io/projected/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-kube-api-access-z45x6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-448ht\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.745978 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z45x6\" (UniqueName: \"kubernetes.io/projected/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-kube-api-access-z45x6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-448ht\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.746264 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-448ht\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.746301 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-448ht\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.752345 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-448ht\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.753117 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-448ht\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.772979 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z45x6\" (UniqueName: \"kubernetes.io/projected/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-kube-api-access-z45x6\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-448ht\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:13 crc kubenswrapper[4676]: I0124 00:37:13.899422 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:14 crc kubenswrapper[4676]: I0124 00:37:14.448754 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht"] Jan 24 00:37:14 crc kubenswrapper[4676]: I0124 00:37:14.501089 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" event={"ID":"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5","Type":"ContainerStarted","Data":"52bbd7242006b7a6c49596bb17ec3e9d18b2e470c6f66a5c43577c74ed4cbaf6"} Jan 24 00:37:14 crc kubenswrapper[4676]: I0124 00:37:14.608785 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:37:14 crc kubenswrapper[4676]: I0124 00:37:14.674192 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:37:14 crc kubenswrapper[4676]: I0124 00:37:14.862183 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-twx92"] Jan 24 00:37:15 crc kubenswrapper[4676]: I0124 00:37:15.511838 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" event={"ID":"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5","Type":"ContainerStarted","Data":"5b3a2fbac507da3687186877115111b773c67d1e6dc327fc4ad3f094369418cb"} Jan 24 00:37:15 crc kubenswrapper[4676]: I0124 00:37:15.544358 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" podStartSLOduration=1.99945335 podStartE2EDuration="2.54432338s" podCreationTimestamp="2026-01-24 00:37:13 +0000 UTC" firstStartedPulling="2026-01-24 00:37:14.456596101 +0000 UTC m=+2018.486567112" lastFinishedPulling="2026-01-24 00:37:15.001466111 +0000 UTC m=+2019.031437142" observedRunningTime="2026-01-24 00:37:15.536738267 +0000 UTC m=+2019.566709288" watchObservedRunningTime="2026-01-24 00:37:15.54432338 +0000 UTC m=+2019.574294431" Jan 24 00:37:16 crc kubenswrapper[4676]: I0124 00:37:16.525347 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-twx92" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" containerName="registry-server" containerID="cri-o://f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b" gracePeriod=2 Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.011988 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.062393 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-catalog-content\") pod \"a3e85f87-765d-44bb-9187-6e5276434b05\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.062496 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxlrz\" (UniqueName: \"kubernetes.io/projected/a3e85f87-765d-44bb-9187-6e5276434b05-kube-api-access-fxlrz\") pod \"a3e85f87-765d-44bb-9187-6e5276434b05\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.062580 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-utilities\") pod \"a3e85f87-765d-44bb-9187-6e5276434b05\" (UID: \"a3e85f87-765d-44bb-9187-6e5276434b05\") " Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.066935 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-utilities" (OuterVolumeSpecName: "utilities") pod "a3e85f87-765d-44bb-9187-6e5276434b05" (UID: "a3e85f87-765d-44bb-9187-6e5276434b05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.077293 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3e85f87-765d-44bb-9187-6e5276434b05-kube-api-access-fxlrz" (OuterVolumeSpecName: "kube-api-access-fxlrz") pod "a3e85f87-765d-44bb-9187-6e5276434b05" (UID: "a3e85f87-765d-44bb-9187-6e5276434b05"). InnerVolumeSpecName "kube-api-access-fxlrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.164241 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.164287 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxlrz\" (UniqueName: \"kubernetes.io/projected/a3e85f87-765d-44bb-9187-6e5276434b05-kube-api-access-fxlrz\") on node \"crc\" DevicePath \"\"" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.190769 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3e85f87-765d-44bb-9187-6e5276434b05" (UID: "a3e85f87-765d-44bb-9187-6e5276434b05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.266358 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3e85f87-765d-44bb-9187-6e5276434b05-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.540270 4676 generic.go:334] "Generic (PLEG): container finished" podID="a3e85f87-765d-44bb-9187-6e5276434b05" containerID="f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b" exitCode=0 Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.540328 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twx92" event={"ID":"a3e85f87-765d-44bb-9187-6e5276434b05","Type":"ContainerDied","Data":"f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b"} Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.540401 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-twx92" event={"ID":"a3e85f87-765d-44bb-9187-6e5276434b05","Type":"ContainerDied","Data":"12f78548df580a50ff02c0e908a128c3171c6cd389e2bfb1762c569bef21f86a"} Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.540433 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-twx92" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.540435 4676 scope.go:117] "RemoveContainer" containerID="f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.579887 4676 scope.go:117] "RemoveContainer" containerID="7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.585128 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-twx92"] Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.604222 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-twx92"] Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.614299 4676 scope.go:117] "RemoveContainer" containerID="c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.650778 4676 scope.go:117] "RemoveContainer" containerID="f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b" Jan 24 00:37:17 crc kubenswrapper[4676]: E0124 00:37:17.651176 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b\": container with ID starting with f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b not found: ID does not exist" containerID="f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.651231 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b"} err="failed to get container status \"f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b\": rpc error: code = NotFound desc = could not find container \"f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b\": container with ID starting with f207c30ece0ab8c19fb55b049c931c60792beebf6fe8e09e87c0dde2cf78a60b not found: ID does not exist" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.651265 4676 scope.go:117] "RemoveContainer" containerID="7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c" Jan 24 00:37:17 crc kubenswrapper[4676]: E0124 00:37:17.651615 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c\": container with ID starting with 7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c not found: ID does not exist" containerID="7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.651650 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c"} err="failed to get container status \"7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c\": rpc error: code = NotFound desc = could not find container \"7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c\": container with ID starting with 7cafafc95566588ce82892ea8b197b1b19666173405b7ac3e5ed41d6f52ba98c not found: ID does not exist" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.651678 4676 scope.go:117] "RemoveContainer" containerID="c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109" Jan 24 00:37:17 crc kubenswrapper[4676]: E0124 00:37:17.651943 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109\": container with ID starting with c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109 not found: ID does not exist" containerID="c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109" Jan 24 00:37:17 crc kubenswrapper[4676]: I0124 00:37:17.651981 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109"} err="failed to get container status \"c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109\": rpc error: code = NotFound desc = could not find container \"c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109\": container with ID starting with c908b71adc9cf1a9bdf7670a770704804db8201b25b1a5fbae60f99ced333109 not found: ID does not exist" Jan 24 00:37:18 crc kubenswrapper[4676]: I0124 00:37:18.274219 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" path="/var/lib/kubelet/pods/a3e85f87-765d-44bb-9187-6e5276434b05/volumes" Jan 24 00:37:25 crc kubenswrapper[4676]: I0124 00:37:25.644889 4676 generic.go:334] "Generic (PLEG): container finished" podID="447c1e1f-d798-4bcc-a8ef-91d4ad5426a5" containerID="5b3a2fbac507da3687186877115111b773c67d1e6dc327fc4ad3f094369418cb" exitCode=0 Jan 24 00:37:25 crc kubenswrapper[4676]: I0124 00:37:25.645580 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" event={"ID":"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5","Type":"ContainerDied","Data":"5b3a2fbac507da3687186877115111b773c67d1e6dc327fc4ad3f094369418cb"} Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.055170 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.163456 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-ssh-key-openstack-edpm-ipam\") pod \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.163581 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z45x6\" (UniqueName: \"kubernetes.io/projected/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-kube-api-access-z45x6\") pod \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.163663 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-inventory\") pod \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\" (UID: \"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5\") " Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.172627 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-kube-api-access-z45x6" (OuterVolumeSpecName: "kube-api-access-z45x6") pod "447c1e1f-d798-4bcc-a8ef-91d4ad5426a5" (UID: "447c1e1f-d798-4bcc-a8ef-91d4ad5426a5"). InnerVolumeSpecName "kube-api-access-z45x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.194621 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "447c1e1f-d798-4bcc-a8ef-91d4ad5426a5" (UID: "447c1e1f-d798-4bcc-a8ef-91d4ad5426a5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.198260 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-inventory" (OuterVolumeSpecName: "inventory") pod "447c1e1f-d798-4bcc-a8ef-91d4ad5426a5" (UID: "447c1e1f-d798-4bcc-a8ef-91d4ad5426a5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.267012 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.267047 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z45x6\" (UniqueName: \"kubernetes.io/projected/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-kube-api-access-z45x6\") on node \"crc\" DevicePath \"\"" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.267060 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/447c1e1f-d798-4bcc-a8ef-91d4ad5426a5-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.666104 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" event={"ID":"447c1e1f-d798-4bcc-a8ef-91d4ad5426a5","Type":"ContainerDied","Data":"52bbd7242006b7a6c49596bb17ec3e9d18b2e470c6f66a5c43577c74ed4cbaf6"} Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.666150 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52bbd7242006b7a6c49596bb17ec3e9d18b2e470c6f66a5c43577c74ed4cbaf6" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.666215 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-448ht" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.807924 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs"] Jan 24 00:37:27 crc kubenswrapper[4676]: E0124 00:37:27.808368 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" containerName="registry-server" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.808406 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" containerName="registry-server" Jan 24 00:37:27 crc kubenswrapper[4676]: E0124 00:37:27.808426 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" containerName="extract-content" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.808435 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" containerName="extract-content" Jan 24 00:37:27 crc kubenswrapper[4676]: E0124 00:37:27.808464 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="447c1e1f-d798-4bcc-a8ef-91d4ad5426a5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.808474 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="447c1e1f-d798-4bcc-a8ef-91d4ad5426a5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:37:27 crc kubenswrapper[4676]: E0124 00:37:27.808501 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" containerName="extract-utilities" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.808511 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" containerName="extract-utilities" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.808727 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3e85f87-765d-44bb-9187-6e5276434b05" containerName="registry-server" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.808779 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="447c1e1f-d798-4bcc-a8ef-91d4ad5426a5" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.809552 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.812706 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.812712 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.813963 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.814519 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.814552 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.814960 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.817017 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.818180 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.818242 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs"] Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.882581 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqhf5\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-kube-api-access-zqhf5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.882673 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.882837 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.882917 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.882986 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.883069 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.883237 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.883327 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.883429 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.883503 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.883613 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.883746 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.883784 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.883873 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985361 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985459 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985501 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985535 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985586 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985630 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985653 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985689 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985743 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqhf5\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-kube-api-access-zqhf5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985772 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985808 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985840 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985868 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.985900 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.990075 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.990935 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.993115 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.993623 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.995539 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.996842 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.997226 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.997303 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:27 crc kubenswrapper[4676]: I0124 00:37:27.998095 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:28 crc kubenswrapper[4676]: I0124 00:37:28.002633 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:28 crc kubenswrapper[4676]: I0124 00:37:28.003274 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:28 crc kubenswrapper[4676]: I0124 00:37:28.005240 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:28 crc kubenswrapper[4676]: I0124 00:37:28.009706 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:28 crc kubenswrapper[4676]: I0124 00:37:28.016600 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqhf5\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-kube-api-access-zqhf5\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:28 crc kubenswrapper[4676]: I0124 00:37:28.171884 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:37:28 crc kubenswrapper[4676]: I0124 00:37:28.567652 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs"] Jan 24 00:37:28 crc kubenswrapper[4676]: I0124 00:37:28.674444 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" event={"ID":"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c","Type":"ContainerStarted","Data":"1de2ad796c2575a2a4e4825e8cb53fd6f1393b20d709f96e27cefc52b380319e"} Jan 24 00:37:29 crc kubenswrapper[4676]: I0124 00:37:29.703041 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" event={"ID":"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c","Type":"ContainerStarted","Data":"ea066811c3fc8f44711798a6a7f1cdd0df2c04ef08ae8fd4743b54536a040057"} Jan 24 00:37:29 crc kubenswrapper[4676]: I0124 00:37:29.741667 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" podStartSLOduration=2.324156073 podStartE2EDuration="2.741646616s" podCreationTimestamp="2026-01-24 00:37:27 +0000 UTC" firstStartedPulling="2026-01-24 00:37:28.56594606 +0000 UTC m=+2032.595917061" lastFinishedPulling="2026-01-24 00:37:28.983436603 +0000 UTC m=+2033.013407604" observedRunningTime="2026-01-24 00:37:29.737976773 +0000 UTC m=+2033.767947814" watchObservedRunningTime="2026-01-24 00:37:29.741646616 +0000 UTC m=+2033.771617637" Jan 24 00:37:39 crc kubenswrapper[4676]: I0124 00:37:39.364102 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:37:39 crc kubenswrapper[4676]: I0124 00:37:39.364844 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:38:09 crc kubenswrapper[4676]: I0124 00:38:09.364240 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:38:09 crc kubenswrapper[4676]: I0124 00:38:09.365075 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:38:14 crc kubenswrapper[4676]: I0124 00:38:14.134109 4676 generic.go:334] "Generic (PLEG): container finished" podID="e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" containerID="ea066811c3fc8f44711798a6a7f1cdd0df2c04ef08ae8fd4743b54536a040057" exitCode=0 Jan 24 00:38:14 crc kubenswrapper[4676]: I0124 00:38:14.134189 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" event={"ID":"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c","Type":"ContainerDied","Data":"ea066811c3fc8f44711798a6a7f1cdd0df2c04ef08ae8fd4743b54536a040057"} Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.607690 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.707120 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-telemetry-combined-ca-bundle\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.707367 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.707510 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-nova-combined-ca-bundle\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.707626 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-bootstrap-combined-ca-bundle\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.707698 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-libvirt-combined-ca-bundle\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.707782 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.707854 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqhf5\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-kube-api-access-zqhf5\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.707930 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-repo-setup-combined-ca-bundle\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.708006 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-inventory\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.708132 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ovn-combined-ca-bundle\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.708211 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ssh-key-openstack-edpm-ipam\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.708287 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.708417 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.708501 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-neutron-metadata-combined-ca-bundle\") pod \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\" (UID: \"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c\") " Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.715988 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.716062 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.717102 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.717240 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.717400 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.719025 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.720436 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.722353 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-kube-api-access-zqhf5" (OuterVolumeSpecName: "kube-api-access-zqhf5") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "kube-api-access-zqhf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.723010 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.725991 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.727248 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.729605 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.742831 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.743196 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-inventory" (OuterVolumeSpecName: "inventory") pod "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" (UID: "e9d2a92f-e22e-44c5-86c5-8b38824e3d4c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810324 4676 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810352 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810363 4676 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810384 4676 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810395 4676 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810404 4676 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810413 4676 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810422 4676 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810432 4676 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810440 4676 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810449 4676 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810458 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqhf5\" (UniqueName: \"kubernetes.io/projected/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-kube-api-access-zqhf5\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810467 4676 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:15 crc kubenswrapper[4676]: I0124 00:38:15.810476 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e9d2a92f-e22e-44c5-86c5-8b38824e3d4c-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.158760 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" event={"ID":"e9d2a92f-e22e-44c5-86c5-8b38824e3d4c","Type":"ContainerDied","Data":"1de2ad796c2575a2a4e4825e8cb53fd6f1393b20d709f96e27cefc52b380319e"} Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.158819 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1de2ad796c2575a2a4e4825e8cb53fd6f1393b20d709f96e27cefc52b380319e" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.158835 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.296455 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv"] Jan 24 00:38:16 crc kubenswrapper[4676]: E0124 00:38:16.296812 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.296832 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.297012 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9d2a92f-e22e-44c5-86c5-8b38824e3d4c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.297573 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.300482 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.300534 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.300584 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.301418 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.301892 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.320261 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv"] Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.422594 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.422700 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.422751 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/55444bfa-a024-4606-aa57-6456c6688e52-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.422780 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n99s\" (UniqueName: \"kubernetes.io/projected/55444bfa-a024-4606-aa57-6456c6688e52-kube-api-access-2n99s\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.422796 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.524969 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.525072 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/55444bfa-a024-4606-aa57-6456c6688e52-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.525137 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n99s\" (UniqueName: \"kubernetes.io/projected/55444bfa-a024-4606-aa57-6456c6688e52-kube-api-access-2n99s\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.525160 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.525292 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.527366 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/55444bfa-a024-4606-aa57-6456c6688e52-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.532355 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.533434 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.538759 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.542131 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n99s\" (UniqueName: \"kubernetes.io/projected/55444bfa-a024-4606-aa57-6456c6688e52-kube-api-access-2n99s\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-527bv\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.615055 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:38:16 crc kubenswrapper[4676]: I0124 00:38:16.985327 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv"] Jan 24 00:38:17 crc kubenswrapper[4676]: I0124 00:38:17.172114 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" event={"ID":"55444bfa-a024-4606-aa57-6456c6688e52","Type":"ContainerStarted","Data":"9ea45bf2e80518f4149bc0176a1374972660dc28f0ba1c11a5d6570b6eca38bf"} Jan 24 00:38:18 crc kubenswrapper[4676]: I0124 00:38:18.185277 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" event={"ID":"55444bfa-a024-4606-aa57-6456c6688e52","Type":"ContainerStarted","Data":"ff5912bb9ffd15ef3b1a9651aa8b86793aa201dc24c7e3f08fe1351670e6d7c0"} Jan 24 00:38:18 crc kubenswrapper[4676]: I0124 00:38:18.218424 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" podStartSLOduration=1.701286338 podStartE2EDuration="2.218356704s" podCreationTimestamp="2026-01-24 00:38:16 +0000 UTC" firstStartedPulling="2026-01-24 00:38:16.999632725 +0000 UTC m=+2081.029603726" lastFinishedPulling="2026-01-24 00:38:17.516703091 +0000 UTC m=+2081.546674092" observedRunningTime="2026-01-24 00:38:18.208661797 +0000 UTC m=+2082.238632838" watchObservedRunningTime="2026-01-24 00:38:18.218356704 +0000 UTC m=+2082.248327735" Jan 24 00:38:39 crc kubenswrapper[4676]: I0124 00:38:39.364435 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:38:39 crc kubenswrapper[4676]: I0124 00:38:39.365175 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:38:39 crc kubenswrapper[4676]: I0124 00:38:39.365251 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:38:39 crc kubenswrapper[4676]: I0124 00:38:39.366604 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"658098ecc7ebeb43955aee4f3317cbcc452c89ec3e2a5ddc24ed196ea90b98d2"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:38:39 crc kubenswrapper[4676]: I0124 00:38:39.366727 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://658098ecc7ebeb43955aee4f3317cbcc452c89ec3e2a5ddc24ed196ea90b98d2" gracePeriod=600 Jan 24 00:38:40 crc kubenswrapper[4676]: I0124 00:38:40.401604 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="658098ecc7ebeb43955aee4f3317cbcc452c89ec3e2a5ddc24ed196ea90b98d2" exitCode=0 Jan 24 00:38:40 crc kubenswrapper[4676]: I0124 00:38:40.401676 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"658098ecc7ebeb43955aee4f3317cbcc452c89ec3e2a5ddc24ed196ea90b98d2"} Jan 24 00:38:40 crc kubenswrapper[4676]: I0124 00:38:40.402100 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac"} Jan 24 00:38:40 crc kubenswrapper[4676]: I0124 00:38:40.402125 4676 scope.go:117] "RemoveContainer" containerID="9c504e3091e16d8ba5c5f912afa13f14ae4119c50ca2e112cebacbce89bb6751" Jan 24 00:39:34 crc kubenswrapper[4676]: I0124 00:39:34.916320 4676 generic.go:334] "Generic (PLEG): container finished" podID="55444bfa-a024-4606-aa57-6456c6688e52" containerID="ff5912bb9ffd15ef3b1a9651aa8b86793aa201dc24c7e3f08fe1351670e6d7c0" exitCode=0 Jan 24 00:39:34 crc kubenswrapper[4676]: I0124 00:39:34.916418 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" event={"ID":"55444bfa-a024-4606-aa57-6456c6688e52","Type":"ContainerDied","Data":"ff5912bb9ffd15ef3b1a9651aa8b86793aa201dc24c7e3f08fe1351670e6d7c0"} Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.443944 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.625891 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-inventory\") pod \"55444bfa-a024-4606-aa57-6456c6688e52\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.626032 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n99s\" (UniqueName: \"kubernetes.io/projected/55444bfa-a024-4606-aa57-6456c6688e52-kube-api-access-2n99s\") pod \"55444bfa-a024-4606-aa57-6456c6688e52\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.627156 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ssh-key-openstack-edpm-ipam\") pod \"55444bfa-a024-4606-aa57-6456c6688e52\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.627249 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ovn-combined-ca-bundle\") pod \"55444bfa-a024-4606-aa57-6456c6688e52\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.627283 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/55444bfa-a024-4606-aa57-6456c6688e52-ovncontroller-config-0\") pod \"55444bfa-a024-4606-aa57-6456c6688e52\" (UID: \"55444bfa-a024-4606-aa57-6456c6688e52\") " Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.632851 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55444bfa-a024-4606-aa57-6456c6688e52-kube-api-access-2n99s" (OuterVolumeSpecName: "kube-api-access-2n99s") pod "55444bfa-a024-4606-aa57-6456c6688e52" (UID: "55444bfa-a024-4606-aa57-6456c6688e52"). InnerVolumeSpecName "kube-api-access-2n99s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.636526 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "55444bfa-a024-4606-aa57-6456c6688e52" (UID: "55444bfa-a024-4606-aa57-6456c6688e52"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.658534 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-inventory" (OuterVolumeSpecName: "inventory") pod "55444bfa-a024-4606-aa57-6456c6688e52" (UID: "55444bfa-a024-4606-aa57-6456c6688e52"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.672938 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55444bfa-a024-4606-aa57-6456c6688e52-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "55444bfa-a024-4606-aa57-6456c6688e52" (UID: "55444bfa-a024-4606-aa57-6456c6688e52"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.672986 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "55444bfa-a024-4606-aa57-6456c6688e52" (UID: "55444bfa-a024-4606-aa57-6456c6688e52"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.729902 4676 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.730066 4676 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/55444bfa-a024-4606-aa57-6456c6688e52-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.730150 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.730353 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n99s\" (UniqueName: \"kubernetes.io/projected/55444bfa-a024-4606-aa57-6456c6688e52-kube-api-access-2n99s\") on node \"crc\" DevicePath \"\"" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.730482 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/55444bfa-a024-4606-aa57-6456c6688e52-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.936570 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" event={"ID":"55444bfa-a024-4606-aa57-6456c6688e52","Type":"ContainerDied","Data":"9ea45bf2e80518f4149bc0176a1374972660dc28f0ba1c11a5d6570b6eca38bf"} Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.936946 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ea45bf2e80518f4149bc0176a1374972660dc28f0ba1c11a5d6570b6eca38bf" Jan 24 00:39:36 crc kubenswrapper[4676]: I0124 00:39:36.936630 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-527bv" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.070187 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7"] Jan 24 00:39:37 crc kubenswrapper[4676]: E0124 00:39:37.070707 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55444bfa-a024-4606-aa57-6456c6688e52" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.070728 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="55444bfa-a024-4606-aa57-6456c6688e52" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.070944 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="55444bfa-a024-4606-aa57-6456c6688e52" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.071833 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.074134 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.074336 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.074583 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.074751 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.075112 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.075844 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.081724 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7"] Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.142197 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.142324 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.142563 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.142921 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.142964 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.143039 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq97x\" (UniqueName: \"kubernetes.io/projected/cba640eb-c65f-46be-af5d-5126418c361a-kube-api-access-qq97x\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.244409 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.244521 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.244545 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.244618 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.244641 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.244662 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq97x\" (UniqueName: \"kubernetes.io/projected/cba640eb-c65f-46be-af5d-5126418c361a-kube-api-access-qq97x\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.250186 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.251853 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.253213 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.254121 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.257073 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.260680 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq97x\" (UniqueName: \"kubernetes.io/projected/cba640eb-c65f-46be-af5d-5126418c361a-kube-api-access-qq97x\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:37 crc kubenswrapper[4676]: I0124 00:39:37.401563 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:39:38 crc kubenswrapper[4676]: I0124 00:39:38.006982 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7"] Jan 24 00:39:38 crc kubenswrapper[4676]: I0124 00:39:38.954929 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" event={"ID":"cba640eb-c65f-46be-af5d-5126418c361a","Type":"ContainerStarted","Data":"768d4cbb4523908822a2ad0486caa07ddbcb8814acacc69c2b351ff7ced7070f"} Jan 24 00:39:38 crc kubenswrapper[4676]: I0124 00:39:38.955178 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" event={"ID":"cba640eb-c65f-46be-af5d-5126418c361a","Type":"ContainerStarted","Data":"87713fa55037d8facd261b161912fac21124b374f9253de8714b996f2b48a4c2"} Jan 24 00:39:38 crc kubenswrapper[4676]: I0124 00:39:38.984363 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" podStartSLOduration=1.433218047 podStartE2EDuration="1.98433532s" podCreationTimestamp="2026-01-24 00:39:37 +0000 UTC" firstStartedPulling="2026-01-24 00:39:38.02554608 +0000 UTC m=+2162.055517101" lastFinishedPulling="2026-01-24 00:39:38.576663373 +0000 UTC m=+2162.606634374" observedRunningTime="2026-01-24 00:39:38.97591325 +0000 UTC m=+2163.005884271" watchObservedRunningTime="2026-01-24 00:39:38.98433532 +0000 UTC m=+2163.014306341" Jan 24 00:40:38 crc kubenswrapper[4676]: I0124 00:40:38.542698 4676 generic.go:334] "Generic (PLEG): container finished" podID="cba640eb-c65f-46be-af5d-5126418c361a" containerID="768d4cbb4523908822a2ad0486caa07ddbcb8814acacc69c2b351ff7ced7070f" exitCode=0 Jan 24 00:40:38 crc kubenswrapper[4676]: I0124 00:40:38.542750 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" event={"ID":"cba640eb-c65f-46be-af5d-5126418c361a","Type":"ContainerDied","Data":"768d4cbb4523908822a2ad0486caa07ddbcb8814acacc69c2b351ff7ced7070f"} Jan 24 00:40:39 crc kubenswrapper[4676]: I0124 00:40:39.364817 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:40:39 crc kubenswrapper[4676]: I0124 00:40:39.365119 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:40:39 crc kubenswrapper[4676]: I0124 00:40:39.980862 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.069171 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-ssh-key-openstack-edpm-ipam\") pod \"cba640eb-c65f-46be-af5d-5126418c361a\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.069302 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-metadata-combined-ca-bundle\") pod \"cba640eb-c65f-46be-af5d-5126418c361a\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.069359 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-ovn-metadata-agent-neutron-config-0\") pod \"cba640eb-c65f-46be-af5d-5126418c361a\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.069409 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-nova-metadata-neutron-config-0\") pod \"cba640eb-c65f-46be-af5d-5126418c361a\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.069447 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-inventory\") pod \"cba640eb-c65f-46be-af5d-5126418c361a\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.069510 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq97x\" (UniqueName: \"kubernetes.io/projected/cba640eb-c65f-46be-af5d-5126418c361a-kube-api-access-qq97x\") pod \"cba640eb-c65f-46be-af5d-5126418c361a\" (UID: \"cba640eb-c65f-46be-af5d-5126418c361a\") " Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.079227 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cba640eb-c65f-46be-af5d-5126418c361a-kube-api-access-qq97x" (OuterVolumeSpecName: "kube-api-access-qq97x") pod "cba640eb-c65f-46be-af5d-5126418c361a" (UID: "cba640eb-c65f-46be-af5d-5126418c361a"). InnerVolumeSpecName "kube-api-access-qq97x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.083791 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "cba640eb-c65f-46be-af5d-5126418c361a" (UID: "cba640eb-c65f-46be-af5d-5126418c361a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.097525 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "cba640eb-c65f-46be-af5d-5126418c361a" (UID: "cba640eb-c65f-46be-af5d-5126418c361a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.098991 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "cba640eb-c65f-46be-af5d-5126418c361a" (UID: "cba640eb-c65f-46be-af5d-5126418c361a"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.102706 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-inventory" (OuterVolumeSpecName: "inventory") pod "cba640eb-c65f-46be-af5d-5126418c361a" (UID: "cba640eb-c65f-46be-af5d-5126418c361a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.114573 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "cba640eb-c65f-46be-af5d-5126418c361a" (UID: "cba640eb-c65f-46be-af5d-5126418c361a"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.172197 4676 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.172273 4676 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.172288 4676 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.172306 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.172320 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq97x\" (UniqueName: \"kubernetes.io/projected/cba640eb-c65f-46be-af5d-5126418c361a-kube-api-access-qq97x\") on node \"crc\" DevicePath \"\"" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.172332 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cba640eb-c65f-46be-af5d-5126418c361a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.565214 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" event={"ID":"cba640eb-c65f-46be-af5d-5126418c361a","Type":"ContainerDied","Data":"87713fa55037d8facd261b161912fac21124b374f9253de8714b996f2b48a4c2"} Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.565249 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.565270 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87713fa55037d8facd261b161912fac21124b374f9253de8714b996f2b48a4c2" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.691845 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99"] Jan 24 00:40:40 crc kubenswrapper[4676]: E0124 00:40:40.692355 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cba640eb-c65f-46be-af5d-5126418c361a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.692408 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="cba640eb-c65f-46be-af5d-5126418c361a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.692793 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="cba640eb-c65f-46be-af5d-5126418c361a" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.694403 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.699309 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.700429 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.700733 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.701972 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.704225 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.715916 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99"] Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.785126 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.785203 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.785393 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.785460 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h6gz\" (UniqueName: \"kubernetes.io/projected/2cafe497-da96-4b39-bec2-1ec54f859303-kube-api-access-9h6gz\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.785585 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.887328 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.887402 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h6gz\" (UniqueName: \"kubernetes.io/projected/2cafe497-da96-4b39-bec2-1ec54f859303-kube-api-access-9h6gz\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.887464 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.887515 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.887555 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.893308 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.893607 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.894660 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.895144 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:40 crc kubenswrapper[4676]: I0124 00:40:40.910856 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h6gz\" (UniqueName: \"kubernetes.io/projected/2cafe497-da96-4b39-bec2-1ec54f859303-kube-api-access-9h6gz\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-ddk99\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:41 crc kubenswrapper[4676]: I0124 00:40:41.018527 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:40:41 crc kubenswrapper[4676]: I0124 00:40:41.607532 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99"] Jan 24 00:40:41 crc kubenswrapper[4676]: I0124 00:40:41.607788 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:40:42 crc kubenswrapper[4676]: I0124 00:40:42.588232 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" event={"ID":"2cafe497-da96-4b39-bec2-1ec54f859303","Type":"ContainerStarted","Data":"9bb06db0d0e1ec2c2b37d9e152501f764e96d00ae7a309b970bb6c0a4f2a01db"} Jan 24 00:40:42 crc kubenswrapper[4676]: I0124 00:40:42.589415 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" event={"ID":"2cafe497-da96-4b39-bec2-1ec54f859303","Type":"ContainerStarted","Data":"1547e9b9c1a3265026b8579e925dfd46811b1b86e0e0cebd473a29138f45bc55"} Jan 24 00:40:42 crc kubenswrapper[4676]: I0124 00:40:42.616616 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" podStartSLOduration=2.072917046 podStartE2EDuration="2.616599349s" podCreationTimestamp="2026-01-24 00:40:40 +0000 UTC" firstStartedPulling="2026-01-24 00:40:41.607542919 +0000 UTC m=+2225.637513920" lastFinishedPulling="2026-01-24 00:40:42.151225222 +0000 UTC m=+2226.181196223" observedRunningTime="2026-01-24 00:40:42.615241307 +0000 UTC m=+2226.645212308" watchObservedRunningTime="2026-01-24 00:40:42.616599349 +0000 UTC m=+2226.646570350" Jan 24 00:41:08 crc kubenswrapper[4676]: I0124 00:41:08.941014 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f4gcx"] Jan 24 00:41:08 crc kubenswrapper[4676]: I0124 00:41:08.946324 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:08 crc kubenswrapper[4676]: I0124 00:41:08.958369 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f4gcx"] Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.016368 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-catalog-content\") pod \"certified-operators-f4gcx\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.016604 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-utilities\") pod \"certified-operators-f4gcx\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.016744 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjcm2\" (UniqueName: \"kubernetes.io/projected/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-kube-api-access-bjcm2\") pod \"certified-operators-f4gcx\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.118240 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjcm2\" (UniqueName: \"kubernetes.io/projected/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-kube-api-access-bjcm2\") pod \"certified-operators-f4gcx\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.118334 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-catalog-content\") pod \"certified-operators-f4gcx\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.118542 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-utilities\") pod \"certified-operators-f4gcx\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.118901 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-catalog-content\") pod \"certified-operators-f4gcx\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.118970 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-utilities\") pod \"certified-operators-f4gcx\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.149022 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjcm2\" (UniqueName: \"kubernetes.io/projected/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-kube-api-access-bjcm2\") pod \"certified-operators-f4gcx\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.279915 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.364770 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.364822 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:41:09 crc kubenswrapper[4676]: I0124 00:41:09.881740 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f4gcx"] Jan 24 00:41:10 crc kubenswrapper[4676]: I0124 00:41:10.902197 4676 generic.go:334] "Generic (PLEG): container finished" podID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerID="737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7" exitCode=0 Jan 24 00:41:10 crc kubenswrapper[4676]: I0124 00:41:10.902249 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4gcx" event={"ID":"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847","Type":"ContainerDied","Data":"737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7"} Jan 24 00:41:10 crc kubenswrapper[4676]: I0124 00:41:10.902570 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4gcx" event={"ID":"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847","Type":"ContainerStarted","Data":"dd0d72b8a2070ae18bd60f3047ac43dfaafa2cc30c6fa2ae4ea63352fa577c6b"} Jan 24 00:41:12 crc kubenswrapper[4676]: I0124 00:41:12.926171 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4gcx" event={"ID":"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847","Type":"ContainerStarted","Data":"5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5"} Jan 24 00:41:13 crc kubenswrapper[4676]: I0124 00:41:13.937680 4676 generic.go:334] "Generic (PLEG): container finished" podID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerID="5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5" exitCode=0 Jan 24 00:41:13 crc kubenswrapper[4676]: I0124 00:41:13.937974 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4gcx" event={"ID":"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847","Type":"ContainerDied","Data":"5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5"} Jan 24 00:41:15 crc kubenswrapper[4676]: I0124 00:41:15.955440 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4gcx" event={"ID":"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847","Type":"ContainerStarted","Data":"b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce"} Jan 24 00:41:15 crc kubenswrapper[4676]: I0124 00:41:15.981273 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f4gcx" podStartSLOduration=4.086522656 podStartE2EDuration="7.981255712s" podCreationTimestamp="2026-01-24 00:41:08 +0000 UTC" firstStartedPulling="2026-01-24 00:41:10.904340695 +0000 UTC m=+2254.934311696" lastFinishedPulling="2026-01-24 00:41:14.799073751 +0000 UTC m=+2258.829044752" observedRunningTime="2026-01-24 00:41:15.979323432 +0000 UTC m=+2260.009294433" watchObservedRunningTime="2026-01-24 00:41:15.981255712 +0000 UTC m=+2260.011226713" Jan 24 00:41:19 crc kubenswrapper[4676]: I0124 00:41:19.280616 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:19 crc kubenswrapper[4676]: I0124 00:41:19.281082 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:19 crc kubenswrapper[4676]: I0124 00:41:19.350672 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.083697 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.140246 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2tgmv"] Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.142032 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.175286 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tgmv"] Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.255578 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-utilities\") pod \"redhat-marketplace-2tgmv\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.255647 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-catalog-content\") pod \"redhat-marketplace-2tgmv\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.255810 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zknhs\" (UniqueName: \"kubernetes.io/projected/5b54a57e-1dba-4552-b03e-94685f89b614-kube-api-access-zknhs\") pod \"redhat-marketplace-2tgmv\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.357358 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zknhs\" (UniqueName: \"kubernetes.io/projected/5b54a57e-1dba-4552-b03e-94685f89b614-kube-api-access-zknhs\") pod \"redhat-marketplace-2tgmv\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.357431 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-utilities\") pod \"redhat-marketplace-2tgmv\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.357460 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-catalog-content\") pod \"redhat-marketplace-2tgmv\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.359073 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-utilities\") pod \"redhat-marketplace-2tgmv\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.359445 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-catalog-content\") pod \"redhat-marketplace-2tgmv\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.389825 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zknhs\" (UniqueName: \"kubernetes.io/projected/5b54a57e-1dba-4552-b03e-94685f89b614-kube-api-access-zknhs\") pod \"redhat-marketplace-2tgmv\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.461663 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:20 crc kubenswrapper[4676]: I0124 00:41:20.954106 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tgmv"] Jan 24 00:41:20 crc kubenswrapper[4676]: W0124 00:41:20.963664 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b54a57e_1dba_4552_b03e_94685f89b614.slice/crio-603cdc7a991c7585c94a502a0ae82308201bad02f44f4c7687a51077bce3e693 WatchSource:0}: Error finding container 603cdc7a991c7585c94a502a0ae82308201bad02f44f4c7687a51077bce3e693: Status 404 returned error can't find the container with id 603cdc7a991c7585c94a502a0ae82308201bad02f44f4c7687a51077bce3e693 Jan 24 00:41:21 crc kubenswrapper[4676]: I0124 00:41:21.004749 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tgmv" event={"ID":"5b54a57e-1dba-4552-b03e-94685f89b614","Type":"ContainerStarted","Data":"603cdc7a991c7585c94a502a0ae82308201bad02f44f4c7687a51077bce3e693"} Jan 24 00:41:21 crc kubenswrapper[4676]: I0124 00:41:21.799813 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f4gcx"] Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.017465 4676 generic.go:334] "Generic (PLEG): container finished" podID="5b54a57e-1dba-4552-b03e-94685f89b614" containerID="c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0" exitCode=0 Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.017552 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tgmv" event={"ID":"5b54a57e-1dba-4552-b03e-94685f89b614","Type":"ContainerDied","Data":"c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0"} Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.018189 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f4gcx" podUID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerName="registry-server" containerID="cri-o://b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce" gracePeriod=2 Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.443248 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.505310 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjcm2\" (UniqueName: \"kubernetes.io/projected/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-kube-api-access-bjcm2\") pod \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.505550 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-catalog-content\") pod \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.505646 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-utilities\") pod \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\" (UID: \"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847\") " Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.506868 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-utilities" (OuterVolumeSpecName: "utilities") pod "96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" (UID: "96dc44b8-b0e8-4f7e-bc80-9b4cf3971847"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.512847 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-kube-api-access-bjcm2" (OuterVolumeSpecName: "kube-api-access-bjcm2") pod "96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" (UID: "96dc44b8-b0e8-4f7e-bc80-9b4cf3971847"). InnerVolumeSpecName "kube-api-access-bjcm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.566970 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" (UID: "96dc44b8-b0e8-4f7e-bc80-9b4cf3971847"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.608658 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.608704 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:41:22 crc kubenswrapper[4676]: I0124 00:41:22.608714 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjcm2\" (UniqueName: \"kubernetes.io/projected/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847-kube-api-access-bjcm2\") on node \"crc\" DevicePath \"\"" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.026291 4676 generic.go:334] "Generic (PLEG): container finished" podID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerID="b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce" exitCode=0 Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.026332 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4gcx" event={"ID":"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847","Type":"ContainerDied","Data":"b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce"} Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.026359 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f4gcx" event={"ID":"96dc44b8-b0e8-4f7e-bc80-9b4cf3971847","Type":"ContainerDied","Data":"dd0d72b8a2070ae18bd60f3047ac43dfaafa2cc30c6fa2ae4ea63352fa577c6b"} Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.026389 4676 scope.go:117] "RemoveContainer" containerID="b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.026367 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f4gcx" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.083575 4676 scope.go:117] "RemoveContainer" containerID="5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.087214 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f4gcx"] Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.098455 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f4gcx"] Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.122411 4676 scope.go:117] "RemoveContainer" containerID="737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.159738 4676 scope.go:117] "RemoveContainer" containerID="b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce" Jan 24 00:41:23 crc kubenswrapper[4676]: E0124 00:41:23.160230 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce\": container with ID starting with b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce not found: ID does not exist" containerID="b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.160279 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce"} err="failed to get container status \"b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce\": rpc error: code = NotFound desc = could not find container \"b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce\": container with ID starting with b429ee368e1d2370a601fe8e4e71726df16a9d3feca5be6e6611ac0b69ecd2ce not found: ID does not exist" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.160316 4676 scope.go:117] "RemoveContainer" containerID="5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5" Jan 24 00:41:23 crc kubenswrapper[4676]: E0124 00:41:23.160699 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5\": container with ID starting with 5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5 not found: ID does not exist" containerID="5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.160748 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5"} err="failed to get container status \"5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5\": rpc error: code = NotFound desc = could not find container \"5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5\": container with ID starting with 5d44871de50f32676b8740620a80ff69be7c3c2e220b2f2348f10f88bfbeede5 not found: ID does not exist" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.160773 4676 scope.go:117] "RemoveContainer" containerID="737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7" Jan 24 00:41:23 crc kubenswrapper[4676]: E0124 00:41:23.161034 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7\": container with ID starting with 737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7 not found: ID does not exist" containerID="737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7" Jan 24 00:41:23 crc kubenswrapper[4676]: I0124 00:41:23.161058 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7"} err="failed to get container status \"737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7\": rpc error: code = NotFound desc = could not find container \"737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7\": container with ID starting with 737d34bc6c9c50d05e277f8c8b483264ff8b1d1d8f9f7474fd3cc837bd2e60f7 not found: ID does not exist" Jan 24 00:41:24 crc kubenswrapper[4676]: I0124 00:41:24.039147 4676 generic.go:334] "Generic (PLEG): container finished" podID="5b54a57e-1dba-4552-b03e-94685f89b614" containerID="a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c" exitCode=0 Jan 24 00:41:24 crc kubenswrapper[4676]: I0124 00:41:24.039557 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tgmv" event={"ID":"5b54a57e-1dba-4552-b03e-94685f89b614","Type":"ContainerDied","Data":"a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c"} Jan 24 00:41:24 crc kubenswrapper[4676]: I0124 00:41:24.271354 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" path="/var/lib/kubelet/pods/96dc44b8-b0e8-4f7e-bc80-9b4cf3971847/volumes" Jan 24 00:41:25 crc kubenswrapper[4676]: I0124 00:41:25.093849 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tgmv" event={"ID":"5b54a57e-1dba-4552-b03e-94685f89b614","Type":"ContainerStarted","Data":"37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be"} Jan 24 00:41:25 crc kubenswrapper[4676]: I0124 00:41:25.124999 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2tgmv" podStartSLOduration=2.52918708 podStartE2EDuration="5.124980312s" podCreationTimestamp="2026-01-24 00:41:20 +0000 UTC" firstStartedPulling="2026-01-24 00:41:22.020820386 +0000 UTC m=+2266.050791397" lastFinishedPulling="2026-01-24 00:41:24.616613628 +0000 UTC m=+2268.646584629" observedRunningTime="2026-01-24 00:41:25.123076922 +0000 UTC m=+2269.153047913" watchObservedRunningTime="2026-01-24 00:41:25.124980312 +0000 UTC m=+2269.154951303" Jan 24 00:41:30 crc kubenswrapper[4676]: I0124 00:41:30.486913 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:30 crc kubenswrapper[4676]: I0124 00:41:30.487810 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:30 crc kubenswrapper[4676]: I0124 00:41:30.549996 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:31 crc kubenswrapper[4676]: I0124 00:41:31.225461 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:31 crc kubenswrapper[4676]: I0124 00:41:31.294010 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tgmv"] Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.177337 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2tgmv" podUID="5b54a57e-1dba-4552-b03e-94685f89b614" containerName="registry-server" containerID="cri-o://37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be" gracePeriod=2 Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.672467 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.760329 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-utilities\") pod \"5b54a57e-1dba-4552-b03e-94685f89b614\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.760544 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-catalog-content\") pod \"5b54a57e-1dba-4552-b03e-94685f89b614\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.760607 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zknhs\" (UniqueName: \"kubernetes.io/projected/5b54a57e-1dba-4552-b03e-94685f89b614-kube-api-access-zknhs\") pod \"5b54a57e-1dba-4552-b03e-94685f89b614\" (UID: \"5b54a57e-1dba-4552-b03e-94685f89b614\") " Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.761263 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-utilities" (OuterVolumeSpecName: "utilities") pod "5b54a57e-1dba-4552-b03e-94685f89b614" (UID: "5b54a57e-1dba-4552-b03e-94685f89b614"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.766631 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b54a57e-1dba-4552-b03e-94685f89b614-kube-api-access-zknhs" (OuterVolumeSpecName: "kube-api-access-zknhs") pod "5b54a57e-1dba-4552-b03e-94685f89b614" (UID: "5b54a57e-1dba-4552-b03e-94685f89b614"). InnerVolumeSpecName "kube-api-access-zknhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.781868 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b54a57e-1dba-4552-b03e-94685f89b614" (UID: "5b54a57e-1dba-4552-b03e-94685f89b614"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.864255 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.864305 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b54a57e-1dba-4552-b03e-94685f89b614-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:41:33 crc kubenswrapper[4676]: I0124 00:41:33.864328 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zknhs\" (UniqueName: \"kubernetes.io/projected/5b54a57e-1dba-4552-b03e-94685f89b614-kube-api-access-zknhs\") on node \"crc\" DevicePath \"\"" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.189503 4676 generic.go:334] "Generic (PLEG): container finished" podID="5b54a57e-1dba-4552-b03e-94685f89b614" containerID="37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be" exitCode=0 Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.189572 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tgmv" event={"ID":"5b54a57e-1dba-4552-b03e-94685f89b614","Type":"ContainerDied","Data":"37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be"} Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.189604 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2tgmv" event={"ID":"5b54a57e-1dba-4552-b03e-94685f89b614","Type":"ContainerDied","Data":"603cdc7a991c7585c94a502a0ae82308201bad02f44f4c7687a51077bce3e693"} Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.189625 4676 scope.go:117] "RemoveContainer" containerID="37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.192300 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2tgmv" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.242824 4676 scope.go:117] "RemoveContainer" containerID="a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.274580 4676 scope.go:117] "RemoveContainer" containerID="c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.276902 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tgmv"] Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.277062 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2tgmv"] Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.332662 4676 scope.go:117] "RemoveContainer" containerID="37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be" Jan 24 00:41:34 crc kubenswrapper[4676]: E0124 00:41:34.333525 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be\": container with ID starting with 37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be not found: ID does not exist" containerID="37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.333578 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be"} err="failed to get container status \"37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be\": rpc error: code = NotFound desc = could not find container \"37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be\": container with ID starting with 37c44ddc8bf250f1ae7443ad12da8e6ead1f79c51b0da873d307014e625972be not found: ID does not exist" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.333611 4676 scope.go:117] "RemoveContainer" containerID="a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c" Jan 24 00:41:34 crc kubenswrapper[4676]: E0124 00:41:34.334087 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c\": container with ID starting with a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c not found: ID does not exist" containerID="a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.334108 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c"} err="failed to get container status \"a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c\": rpc error: code = NotFound desc = could not find container \"a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c\": container with ID starting with a8722a93ad722ca622f815f683251c0b2fa480044fc8a3a7f7bbcb10216c506c not found: ID does not exist" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.334122 4676 scope.go:117] "RemoveContainer" containerID="c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0" Jan 24 00:41:34 crc kubenswrapper[4676]: E0124 00:41:34.334524 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0\": container with ID starting with c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0 not found: ID does not exist" containerID="c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0" Jan 24 00:41:34 crc kubenswrapper[4676]: I0124 00:41:34.334547 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0"} err="failed to get container status \"c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0\": rpc error: code = NotFound desc = could not find container \"c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0\": container with ID starting with c2701b192e86de32a0ec79475d981d5c7c6144cfecdced05bb10231d98dfc4e0 not found: ID does not exist" Jan 24 00:41:36 crc kubenswrapper[4676]: I0124 00:41:36.291772 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b54a57e-1dba-4552-b03e-94685f89b614" path="/var/lib/kubelet/pods/5b54a57e-1dba-4552-b03e-94685f89b614/volumes" Jan 24 00:41:39 crc kubenswrapper[4676]: I0124 00:41:39.364595 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:41:39 crc kubenswrapper[4676]: I0124 00:41:39.364926 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:41:39 crc kubenswrapper[4676]: I0124 00:41:39.364964 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:41:39 crc kubenswrapper[4676]: I0124 00:41:39.365777 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:41:39 crc kubenswrapper[4676]: I0124 00:41:39.365832 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" gracePeriod=600 Jan 24 00:41:39 crc kubenswrapper[4676]: E0124 00:41:39.502644 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:41:40 crc kubenswrapper[4676]: I0124 00:41:40.288272 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" exitCode=0 Jan 24 00:41:40 crc kubenswrapper[4676]: I0124 00:41:40.289575 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac"} Jan 24 00:41:40 crc kubenswrapper[4676]: I0124 00:41:40.289655 4676 scope.go:117] "RemoveContainer" containerID="658098ecc7ebeb43955aee4f3317cbcc452c89ec3e2a5ddc24ed196ea90b98d2" Jan 24 00:41:40 crc kubenswrapper[4676]: I0124 00:41:40.290610 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:41:40 crc kubenswrapper[4676]: E0124 00:41:40.291063 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:41:54 crc kubenswrapper[4676]: I0124 00:41:54.255916 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:41:54 crc kubenswrapper[4676]: E0124 00:41:54.256788 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:42:05 crc kubenswrapper[4676]: I0124 00:42:05.256580 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:42:05 crc kubenswrapper[4676]: E0124 00:42:05.257213 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:42:18 crc kubenswrapper[4676]: I0124 00:42:18.257945 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:42:18 crc kubenswrapper[4676]: E0124 00:42:18.259346 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:42:32 crc kubenswrapper[4676]: I0124 00:42:32.258006 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:42:32 crc kubenswrapper[4676]: E0124 00:42:32.259183 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:42:45 crc kubenswrapper[4676]: I0124 00:42:45.257430 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:42:45 crc kubenswrapper[4676]: E0124 00:42:45.258591 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:42:57 crc kubenswrapper[4676]: I0124 00:42:57.255989 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:42:57 crc kubenswrapper[4676]: E0124 00:42:57.257142 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:43:10 crc kubenswrapper[4676]: I0124 00:43:10.256268 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:43:10 crc kubenswrapper[4676]: E0124 00:43:10.257409 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:43:21 crc kubenswrapper[4676]: I0124 00:43:21.257436 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:43:21 crc kubenswrapper[4676]: E0124 00:43:21.260989 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:43:36 crc kubenswrapper[4676]: I0124 00:43:36.265981 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:43:36 crc kubenswrapper[4676]: E0124 00:43:36.266732 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:43:47 crc kubenswrapper[4676]: I0124 00:43:47.256058 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:43:47 crc kubenswrapper[4676]: E0124 00:43:47.256994 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:43:58 crc kubenswrapper[4676]: I0124 00:43:58.255708 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:43:58 crc kubenswrapper[4676]: E0124 00:43:58.256522 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:44:09 crc kubenswrapper[4676]: I0124 00:44:09.255749 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:44:09 crc kubenswrapper[4676]: E0124 00:44:09.256289 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:44:20 crc kubenswrapper[4676]: I0124 00:44:20.256126 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:44:20 crc kubenswrapper[4676]: E0124 00:44:20.257022 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:44:35 crc kubenswrapper[4676]: I0124 00:44:35.260075 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:44:35 crc kubenswrapper[4676]: E0124 00:44:35.261189 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:44:47 crc kubenswrapper[4676]: I0124 00:44:47.255993 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:44:47 crc kubenswrapper[4676]: E0124 00:44:47.256786 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:44:59 crc kubenswrapper[4676]: I0124 00:44:59.257168 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:44:59 crc kubenswrapper[4676]: E0124 00:44:59.258165 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.167979 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt"] Jan 24 00:45:00 crc kubenswrapper[4676]: E0124 00:45:00.168923 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54a57e-1dba-4552-b03e-94685f89b614" containerName="registry-server" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.168970 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54a57e-1dba-4552-b03e-94685f89b614" containerName="registry-server" Jan 24 00:45:00 crc kubenswrapper[4676]: E0124 00:45:00.169032 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerName="registry-server" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.169052 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerName="registry-server" Jan 24 00:45:00 crc kubenswrapper[4676]: E0124 00:45:00.169075 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54a57e-1dba-4552-b03e-94685f89b614" containerName="extract-content" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.169092 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54a57e-1dba-4552-b03e-94685f89b614" containerName="extract-content" Jan 24 00:45:00 crc kubenswrapper[4676]: E0124 00:45:00.169117 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerName="extract-content" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.169132 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerName="extract-content" Jan 24 00:45:00 crc kubenswrapper[4676]: E0124 00:45:00.169169 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b54a57e-1dba-4552-b03e-94685f89b614" containerName="extract-utilities" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.169186 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b54a57e-1dba-4552-b03e-94685f89b614" containerName="extract-utilities" Jan 24 00:45:00 crc kubenswrapper[4676]: E0124 00:45:00.169239 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerName="extract-utilities" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.169258 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerName="extract-utilities" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.169727 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b54a57e-1dba-4552-b03e-94685f89b614" containerName="registry-server" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.169783 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="96dc44b8-b0e8-4f7e-bc80-9b4cf3971847" containerName="registry-server" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.171252 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.175480 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.188082 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt"] Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.189522 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.267226 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5lfg\" (UniqueName: \"kubernetes.io/projected/8d2829b1-9d20-469f-8c25-475f980b00f1-kube-api-access-p5lfg\") pod \"collect-profiles-29486925-f7sqt\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.267461 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d2829b1-9d20-469f-8c25-475f980b00f1-secret-volume\") pod \"collect-profiles-29486925-f7sqt\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.267491 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d2829b1-9d20-469f-8c25-475f980b00f1-config-volume\") pod \"collect-profiles-29486925-f7sqt\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.369434 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5lfg\" (UniqueName: \"kubernetes.io/projected/8d2829b1-9d20-469f-8c25-475f980b00f1-kube-api-access-p5lfg\") pod \"collect-profiles-29486925-f7sqt\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.369476 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d2829b1-9d20-469f-8c25-475f980b00f1-secret-volume\") pod \"collect-profiles-29486925-f7sqt\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.369505 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d2829b1-9d20-469f-8c25-475f980b00f1-config-volume\") pod \"collect-profiles-29486925-f7sqt\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.370272 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d2829b1-9d20-469f-8c25-475f980b00f1-config-volume\") pod \"collect-profiles-29486925-f7sqt\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.378203 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d2829b1-9d20-469f-8c25-475f980b00f1-secret-volume\") pod \"collect-profiles-29486925-f7sqt\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.390156 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5lfg\" (UniqueName: \"kubernetes.io/projected/8d2829b1-9d20-469f-8c25-475f980b00f1-kube-api-access-p5lfg\") pod \"collect-profiles-29486925-f7sqt\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.491582 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:00 crc kubenswrapper[4676]: I0124 00:45:00.960276 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt"] Jan 24 00:45:01 crc kubenswrapper[4676]: I0124 00:45:01.546186 4676 generic.go:334] "Generic (PLEG): container finished" podID="8d2829b1-9d20-469f-8c25-475f980b00f1" containerID="8a48ddd1d6c47259792d6571cdd4d4b680b060204c9a096dda74141930a9ad12" exitCode=0 Jan 24 00:45:01 crc kubenswrapper[4676]: I0124 00:45:01.546299 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" event={"ID":"8d2829b1-9d20-469f-8c25-475f980b00f1","Type":"ContainerDied","Data":"8a48ddd1d6c47259792d6571cdd4d4b680b060204c9a096dda74141930a9ad12"} Jan 24 00:45:01 crc kubenswrapper[4676]: I0124 00:45:01.547539 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" event={"ID":"8d2829b1-9d20-469f-8c25-475f980b00f1","Type":"ContainerStarted","Data":"960fa06c3628fbca73cfcc44d608c7a1780cd8a341d3c0e7d4ffb4c586e5c67d"} Jan 24 00:45:02 crc kubenswrapper[4676]: I0124 00:45:02.891658 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.017137 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d2829b1-9d20-469f-8c25-475f980b00f1-config-volume\") pod \"8d2829b1-9d20-469f-8c25-475f980b00f1\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.017612 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d2829b1-9d20-469f-8c25-475f980b00f1-secret-volume\") pod \"8d2829b1-9d20-469f-8c25-475f980b00f1\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.018057 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d2829b1-9d20-469f-8c25-475f980b00f1-config-volume" (OuterVolumeSpecName: "config-volume") pod "8d2829b1-9d20-469f-8c25-475f980b00f1" (UID: "8d2829b1-9d20-469f-8c25-475f980b00f1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.018188 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5lfg\" (UniqueName: \"kubernetes.io/projected/8d2829b1-9d20-469f-8c25-475f980b00f1-kube-api-access-p5lfg\") pod \"8d2829b1-9d20-469f-8c25-475f980b00f1\" (UID: \"8d2829b1-9d20-469f-8c25-475f980b00f1\") " Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.019081 4676 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8d2829b1-9d20-469f-8c25-475f980b00f1-config-volume\") on node \"crc\" DevicePath \"\"" Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.023010 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d2829b1-9d20-469f-8c25-475f980b00f1-kube-api-access-p5lfg" (OuterVolumeSpecName: "kube-api-access-p5lfg") pod "8d2829b1-9d20-469f-8c25-475f980b00f1" (UID: "8d2829b1-9d20-469f-8c25-475f980b00f1"). InnerVolumeSpecName "kube-api-access-p5lfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.024180 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d2829b1-9d20-469f-8c25-475f980b00f1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8d2829b1-9d20-469f-8c25-475f980b00f1" (UID: "8d2829b1-9d20-469f-8c25-475f980b00f1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.120923 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5lfg\" (UniqueName: \"kubernetes.io/projected/8d2829b1-9d20-469f-8c25-475f980b00f1-kube-api-access-p5lfg\") on node \"crc\" DevicePath \"\"" Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.121128 4676 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8d2829b1-9d20-469f-8c25-475f980b00f1-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.570444 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" event={"ID":"8d2829b1-9d20-469f-8c25-475f980b00f1","Type":"ContainerDied","Data":"960fa06c3628fbca73cfcc44d608c7a1780cd8a341d3c0e7d4ffb4c586e5c67d"} Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.570501 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="960fa06c3628fbca73cfcc44d608c7a1780cd8a341d3c0e7d4ffb4c586e5c67d" Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.570595 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486925-f7sqt" Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.967547 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg"] Jan 24 00:45:03 crc kubenswrapper[4676]: I0124 00:45:03.972423 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486880-7srrg"] Jan 24 00:45:04 crc kubenswrapper[4676]: I0124 00:45:04.280060 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eabf562-d289-4685-8ee5-ed1525930d19" path="/var/lib/kubelet/pods/1eabf562-d289-4685-8ee5-ed1525930d19/volumes" Jan 24 00:45:12 crc kubenswrapper[4676]: I0124 00:45:12.255793 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:45:12 crc kubenswrapper[4676]: E0124 00:45:12.256898 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:45:24 crc kubenswrapper[4676]: I0124 00:45:24.256967 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:45:24 crc kubenswrapper[4676]: E0124 00:45:24.257993 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:45:36 crc kubenswrapper[4676]: I0124 00:45:36.603291 4676 generic.go:334] "Generic (PLEG): container finished" podID="2cafe497-da96-4b39-bec2-1ec54f859303" containerID="9bb06db0d0e1ec2c2b37d9e152501f764e96d00ae7a309b970bb6c0a4f2a01db" exitCode=0 Jan 24 00:45:36 crc kubenswrapper[4676]: I0124 00:45:36.603603 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" event={"ID":"2cafe497-da96-4b39-bec2-1ec54f859303","Type":"ContainerDied","Data":"9bb06db0d0e1ec2c2b37d9e152501f764e96d00ae7a309b970bb6c0a4f2a01db"} Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.043551 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.152884 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-combined-ca-bundle\") pod \"2cafe497-da96-4b39-bec2-1ec54f859303\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.153319 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h6gz\" (UniqueName: \"kubernetes.io/projected/2cafe497-da96-4b39-bec2-1ec54f859303-kube-api-access-9h6gz\") pod \"2cafe497-da96-4b39-bec2-1ec54f859303\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.153375 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-ssh-key-openstack-edpm-ipam\") pod \"2cafe497-da96-4b39-bec2-1ec54f859303\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.153435 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-secret-0\") pod \"2cafe497-da96-4b39-bec2-1ec54f859303\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.153636 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-inventory\") pod \"2cafe497-da96-4b39-bec2-1ec54f859303\" (UID: \"2cafe497-da96-4b39-bec2-1ec54f859303\") " Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.162665 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2cafe497-da96-4b39-bec2-1ec54f859303" (UID: "2cafe497-da96-4b39-bec2-1ec54f859303"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.163854 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cafe497-da96-4b39-bec2-1ec54f859303-kube-api-access-9h6gz" (OuterVolumeSpecName: "kube-api-access-9h6gz") pod "2cafe497-da96-4b39-bec2-1ec54f859303" (UID: "2cafe497-da96-4b39-bec2-1ec54f859303"). InnerVolumeSpecName "kube-api-access-9h6gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.186874 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-inventory" (OuterVolumeSpecName: "inventory") pod "2cafe497-da96-4b39-bec2-1ec54f859303" (UID: "2cafe497-da96-4b39-bec2-1ec54f859303"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.190115 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "2cafe497-da96-4b39-bec2-1ec54f859303" (UID: "2cafe497-da96-4b39-bec2-1ec54f859303"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.199087 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2cafe497-da96-4b39-bec2-1ec54f859303" (UID: "2cafe497-da96-4b39-bec2-1ec54f859303"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.256592 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.256644 4676 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.256663 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.256686 4676 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cafe497-da96-4b39-bec2-1ec54f859303-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.256706 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h6gz\" (UniqueName: \"kubernetes.io/projected/2cafe497-da96-4b39-bec2-1ec54f859303-kube-api-access-9h6gz\") on node \"crc\" DevicePath \"\"" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.631123 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" event={"ID":"2cafe497-da96-4b39-bec2-1ec54f859303","Type":"ContainerDied","Data":"1547e9b9c1a3265026b8579e925dfd46811b1b86e0e0cebd473a29138f45bc55"} Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.631176 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1547e9b9c1a3265026b8579e925dfd46811b1b86e0e0cebd473a29138f45bc55" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.631551 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-ddk99" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.737156 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g"] Jan 24 00:45:38 crc kubenswrapper[4676]: E0124 00:45:38.737524 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cafe497-da96-4b39-bec2-1ec54f859303" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.737541 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cafe497-da96-4b39-bec2-1ec54f859303" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 24 00:45:38 crc kubenswrapper[4676]: E0124 00:45:38.737549 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d2829b1-9d20-469f-8c25-475f980b00f1" containerName="collect-profiles" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.737556 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d2829b1-9d20-469f-8c25-475f980b00f1" containerName="collect-profiles" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.737714 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d2829b1-9d20-469f-8c25-475f980b00f1" containerName="collect-profiles" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.737732 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cafe497-da96-4b39-bec2-1ec54f859303" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.738253 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.741058 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.741482 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.741907 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.742212 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.743338 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.743818 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.753679 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.765210 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g"] Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.876910 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.876959 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6689\" (UniqueName: \"kubernetes.io/projected/75cae94d-b818-4f92-b42d-fd8cec63a657-kube-api-access-g6689\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.876981 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.877006 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.877108 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.877178 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.877199 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.877403 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.877468 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.979752 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.979806 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.979888 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.979929 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6689\" (UniqueName: \"kubernetes.io/projected/75cae94d-b818-4f92-b42d-fd8cec63a657-kube-api-access-g6689\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.979946 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.979970 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.979998 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.980058 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.980075 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.981310 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.985666 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.985822 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:38 crc kubenswrapper[4676]: I0124 00:45:38.993823 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:39 crc kubenswrapper[4676]: I0124 00:45:39.000153 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:39 crc kubenswrapper[4676]: I0124 00:45:39.001714 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:39 crc kubenswrapper[4676]: I0124 00:45:39.002175 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:39 crc kubenswrapper[4676]: I0124 00:45:39.006896 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:39 crc kubenswrapper[4676]: I0124 00:45:39.010057 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6689\" (UniqueName: \"kubernetes.io/projected/75cae94d-b818-4f92-b42d-fd8cec63a657-kube-api-access-g6689\") pod \"nova-edpm-deployment-openstack-edpm-ipam-5cl2g\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:39 crc kubenswrapper[4676]: I0124 00:45:39.059268 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:45:39 crc kubenswrapper[4676]: I0124 00:45:39.255830 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:45:39 crc kubenswrapper[4676]: E0124 00:45:39.256505 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:45:39 crc kubenswrapper[4676]: I0124 00:45:39.447889 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g"] Jan 24 00:45:39 crc kubenswrapper[4676]: W0124 00:45:39.460572 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75cae94d_b818_4f92_b42d_fd8cec63a657.slice/crio-7a1a57426601586865e4d5bc7b431f54c628df289eb8caca5bd5436e21faee21 WatchSource:0}: Error finding container 7a1a57426601586865e4d5bc7b431f54c628df289eb8caca5bd5436e21faee21: Status 404 returned error can't find the container with id 7a1a57426601586865e4d5bc7b431f54c628df289eb8caca5bd5436e21faee21 Jan 24 00:45:39 crc kubenswrapper[4676]: I0124 00:45:39.641756 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" event={"ID":"75cae94d-b818-4f92-b42d-fd8cec63a657","Type":"ContainerStarted","Data":"7a1a57426601586865e4d5bc7b431f54c628df289eb8caca5bd5436e21faee21"} Jan 24 00:45:40 crc kubenswrapper[4676]: I0124 00:45:40.654128 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" event={"ID":"75cae94d-b818-4f92-b42d-fd8cec63a657","Type":"ContainerStarted","Data":"9049acca3eaa19729de3abedb765017e8466c178a888c60c1ddeb96f4f0c0556"} Jan 24 00:45:40 crc kubenswrapper[4676]: I0124 00:45:40.704519 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" podStartSLOduration=2.173199826 podStartE2EDuration="2.704502101s" podCreationTimestamp="2026-01-24 00:45:38 +0000 UTC" firstStartedPulling="2026-01-24 00:45:39.463773204 +0000 UTC m=+2523.493744215" lastFinishedPulling="2026-01-24 00:45:39.995075489 +0000 UTC m=+2524.025046490" observedRunningTime="2026-01-24 00:45:40.69861277 +0000 UTC m=+2524.728583801" watchObservedRunningTime="2026-01-24 00:45:40.704502101 +0000 UTC m=+2524.734473102" Jan 24 00:45:42 crc kubenswrapper[4676]: I0124 00:45:42.712727 4676 scope.go:117] "RemoveContainer" containerID="572cf18d919a142df3a87fe72606c221700742f64f3b5868d79f863d9965c925" Jan 24 00:45:52 crc kubenswrapper[4676]: I0124 00:45:52.256253 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:45:52 crc kubenswrapper[4676]: E0124 00:45:52.257184 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.295535 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9gbql"] Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.298415 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.321267 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9gbql"] Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.442734 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpsct\" (UniqueName: \"kubernetes.io/projected/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-kube-api-access-mpsct\") pod \"community-operators-9gbql\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.442800 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-utilities\") pod \"community-operators-9gbql\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.442821 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-catalog-content\") pod \"community-operators-9gbql\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.544730 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpsct\" (UniqueName: \"kubernetes.io/projected/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-kube-api-access-mpsct\") pod \"community-operators-9gbql\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.544793 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-utilities\") pod \"community-operators-9gbql\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.544825 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-catalog-content\") pod \"community-operators-9gbql\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.545328 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-utilities\") pod \"community-operators-9gbql\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.545486 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-catalog-content\") pod \"community-operators-9gbql\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.564689 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpsct\" (UniqueName: \"kubernetes.io/projected/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-kube-api-access-mpsct\") pod \"community-operators-9gbql\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:00 crc kubenswrapper[4676]: I0124 00:46:00.621558 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:01 crc kubenswrapper[4676]: I0124 00:46:01.156857 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9gbql"] Jan 24 00:46:01 crc kubenswrapper[4676]: I0124 00:46:01.879409 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gbql" event={"ID":"20acbbf0-6ea7-4138-9b68-4ab7b31071e5","Type":"ContainerDied","Data":"e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e"} Jan 24 00:46:01 crc kubenswrapper[4676]: I0124 00:46:01.879347 4676 generic.go:334] "Generic (PLEG): container finished" podID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerID="e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e" exitCode=0 Jan 24 00:46:01 crc kubenswrapper[4676]: I0124 00:46:01.879493 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gbql" event={"ID":"20acbbf0-6ea7-4138-9b68-4ab7b31071e5","Type":"ContainerStarted","Data":"2f61a2bc6b2a611ed6a3f9e74e1dcca47d070d91d158983ca73e39445419936f"} Jan 24 00:46:01 crc kubenswrapper[4676]: I0124 00:46:01.882367 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:46:02 crc kubenswrapper[4676]: I0124 00:46:02.890356 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gbql" event={"ID":"20acbbf0-6ea7-4138-9b68-4ab7b31071e5","Type":"ContainerStarted","Data":"01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3"} Jan 24 00:46:04 crc kubenswrapper[4676]: I0124 00:46:04.011523 4676 generic.go:334] "Generic (PLEG): container finished" podID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerID="01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3" exitCode=0 Jan 24 00:46:04 crc kubenswrapper[4676]: I0124 00:46:04.011571 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gbql" event={"ID":"20acbbf0-6ea7-4138-9b68-4ab7b31071e5","Type":"ContainerDied","Data":"01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3"} Jan 24 00:46:05 crc kubenswrapper[4676]: I0124 00:46:05.022538 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gbql" event={"ID":"20acbbf0-6ea7-4138-9b68-4ab7b31071e5","Type":"ContainerStarted","Data":"727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893"} Jan 24 00:46:05 crc kubenswrapper[4676]: I0124 00:46:05.075537 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9gbql" podStartSLOduration=2.556075646 podStartE2EDuration="5.075512843s" podCreationTimestamp="2026-01-24 00:46:00 +0000 UTC" firstStartedPulling="2026-01-24 00:46:01.881860098 +0000 UTC m=+2545.911831099" lastFinishedPulling="2026-01-24 00:46:04.401297295 +0000 UTC m=+2548.431268296" observedRunningTime="2026-01-24 00:46:05.069645802 +0000 UTC m=+2549.099616843" watchObservedRunningTime="2026-01-24 00:46:05.075512843 +0000 UTC m=+2549.105483854" Jan 24 00:46:06 crc kubenswrapper[4676]: I0124 00:46:06.261706 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:46:06 crc kubenswrapper[4676]: E0124 00:46:06.262178 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:46:10 crc kubenswrapper[4676]: I0124 00:46:10.622479 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:10 crc kubenswrapper[4676]: I0124 00:46:10.622866 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:10 crc kubenswrapper[4676]: I0124 00:46:10.698500 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:11 crc kubenswrapper[4676]: I0124 00:46:11.136502 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:11 crc kubenswrapper[4676]: I0124 00:46:11.202485 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9gbql"] Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.107874 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9gbql" podUID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerName="registry-server" containerID="cri-o://727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893" gracePeriod=2 Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.573979 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.707072 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpsct\" (UniqueName: \"kubernetes.io/projected/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-kube-api-access-mpsct\") pod \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.707465 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-utilities\") pod \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.707558 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-catalog-content\") pod \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\" (UID: \"20acbbf0-6ea7-4138-9b68-4ab7b31071e5\") " Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.710217 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-utilities" (OuterVolumeSpecName: "utilities") pod "20acbbf0-6ea7-4138-9b68-4ab7b31071e5" (UID: "20acbbf0-6ea7-4138-9b68-4ab7b31071e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.717684 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-kube-api-access-mpsct" (OuterVolumeSpecName: "kube-api-access-mpsct") pod "20acbbf0-6ea7-4138-9b68-4ab7b31071e5" (UID: "20acbbf0-6ea7-4138-9b68-4ab7b31071e5"). InnerVolumeSpecName "kube-api-access-mpsct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.791147 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20acbbf0-6ea7-4138-9b68-4ab7b31071e5" (UID: "20acbbf0-6ea7-4138-9b68-4ab7b31071e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.809966 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpsct\" (UniqueName: \"kubernetes.io/projected/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-kube-api-access-mpsct\") on node \"crc\" DevicePath \"\"" Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.810003 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:46:13 crc kubenswrapper[4676]: I0124 00:46:13.810015 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20acbbf0-6ea7-4138-9b68-4ab7b31071e5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.119669 4676 generic.go:334] "Generic (PLEG): container finished" podID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerID="727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893" exitCode=0 Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.119733 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9gbql" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.119750 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gbql" event={"ID":"20acbbf0-6ea7-4138-9b68-4ab7b31071e5","Type":"ContainerDied","Data":"727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893"} Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.119812 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9gbql" event={"ID":"20acbbf0-6ea7-4138-9b68-4ab7b31071e5","Type":"ContainerDied","Data":"2f61a2bc6b2a611ed6a3f9e74e1dcca47d070d91d158983ca73e39445419936f"} Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.119836 4676 scope.go:117] "RemoveContainer" containerID="727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.152515 4676 scope.go:117] "RemoveContainer" containerID="01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.189423 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9gbql"] Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.220767 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9gbql"] Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.224098 4676 scope.go:117] "RemoveContainer" containerID="e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.251221 4676 scope.go:117] "RemoveContainer" containerID="727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893" Jan 24 00:46:14 crc kubenswrapper[4676]: E0124 00:46:14.252803 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893\": container with ID starting with 727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893 not found: ID does not exist" containerID="727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.252833 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893"} err="failed to get container status \"727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893\": rpc error: code = NotFound desc = could not find container \"727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893\": container with ID starting with 727704ff4f53ba7b13f4b78ff613c85e8daf69033f4490df8504f6bfe1a82893 not found: ID does not exist" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.252858 4676 scope.go:117] "RemoveContainer" containerID="01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3" Jan 24 00:46:14 crc kubenswrapper[4676]: E0124 00:46:14.253211 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3\": container with ID starting with 01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3 not found: ID does not exist" containerID="01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.253233 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3"} err="failed to get container status \"01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3\": rpc error: code = NotFound desc = could not find container \"01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3\": container with ID starting with 01b3ed872259c694bfc2d43f160db2bacf44f5bfd0e78056095d41e8770fd6f3 not found: ID does not exist" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.253247 4676 scope.go:117] "RemoveContainer" containerID="e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e" Jan 24 00:46:14 crc kubenswrapper[4676]: E0124 00:46:14.253542 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e\": container with ID starting with e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e not found: ID does not exist" containerID="e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.253560 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e"} err="failed to get container status \"e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e\": rpc error: code = NotFound desc = could not find container \"e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e\": container with ID starting with e8ecd8886191eaccb5939178d7358f3825d6ba6249f300176017ef377067634e not found: ID does not exist" Jan 24 00:46:14 crc kubenswrapper[4676]: I0124 00:46:14.266904 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" path="/var/lib/kubelet/pods/20acbbf0-6ea7-4138-9b68-4ab7b31071e5/volumes" Jan 24 00:46:19 crc kubenswrapper[4676]: I0124 00:46:19.257352 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:46:19 crc kubenswrapper[4676]: E0124 00:46:19.258344 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:46:30 crc kubenswrapper[4676]: I0124 00:46:30.257061 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:46:30 crc kubenswrapper[4676]: E0124 00:46:30.258153 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:46:42 crc kubenswrapper[4676]: I0124 00:46:42.256155 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:46:43 crc kubenswrapper[4676]: I0124 00:46:43.423151 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"50556dab9c423476f3e768a6107a078181f27e372f40fe19e80cac55ab001315"} Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.447729 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8bhxt"] Jan 24 00:47:50 crc kubenswrapper[4676]: E0124 00:47:50.448611 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerName="extract-utilities" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.448628 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerName="extract-utilities" Jan 24 00:47:50 crc kubenswrapper[4676]: E0124 00:47:50.448663 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerName="extract-content" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.448673 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerName="extract-content" Jan 24 00:47:50 crc kubenswrapper[4676]: E0124 00:47:50.448693 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerName="registry-server" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.448702 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerName="registry-server" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.448916 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="20acbbf0-6ea7-4138-9b68-4ab7b31071e5" containerName="registry-server" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.453491 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.472222 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8bhxt"] Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.589356 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-utilities\") pod \"redhat-operators-8bhxt\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.589432 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-catalog-content\") pod \"redhat-operators-8bhxt\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.589846 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpkj4\" (UniqueName: \"kubernetes.io/projected/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-kube-api-access-rpkj4\") pod \"redhat-operators-8bhxt\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.691831 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpkj4\" (UniqueName: \"kubernetes.io/projected/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-kube-api-access-rpkj4\") pod \"redhat-operators-8bhxt\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.691960 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-utilities\") pod \"redhat-operators-8bhxt\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.692007 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-catalog-content\") pod \"redhat-operators-8bhxt\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.692705 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-catalog-content\") pod \"redhat-operators-8bhxt\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.692831 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-utilities\") pod \"redhat-operators-8bhxt\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.716734 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpkj4\" (UniqueName: \"kubernetes.io/projected/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-kube-api-access-rpkj4\") pod \"redhat-operators-8bhxt\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:50 crc kubenswrapper[4676]: I0124 00:47:50.773436 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:47:51 crc kubenswrapper[4676]: I0124 00:47:51.279760 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8bhxt"] Jan 24 00:47:52 crc kubenswrapper[4676]: I0124 00:47:52.116049 4676 generic.go:334] "Generic (PLEG): container finished" podID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerID="b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08" exitCode=0 Jan 24 00:47:52 crc kubenswrapper[4676]: I0124 00:47:52.116317 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bhxt" event={"ID":"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846","Type":"ContainerDied","Data":"b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08"} Jan 24 00:47:52 crc kubenswrapper[4676]: I0124 00:47:52.116343 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bhxt" event={"ID":"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846","Type":"ContainerStarted","Data":"b2d4b4da885e3e4644df949bcafb64ca58f900ed27867e562ecef9983b65ab9e"} Jan 24 00:47:54 crc kubenswrapper[4676]: I0124 00:47:54.132205 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bhxt" event={"ID":"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846","Type":"ContainerStarted","Data":"2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7"} Jan 24 00:47:57 crc kubenswrapper[4676]: I0124 00:47:57.157639 4676 generic.go:334] "Generic (PLEG): container finished" podID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerID="2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7" exitCode=0 Jan 24 00:47:57 crc kubenswrapper[4676]: I0124 00:47:57.157688 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bhxt" event={"ID":"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846","Type":"ContainerDied","Data":"2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7"} Jan 24 00:47:58 crc kubenswrapper[4676]: I0124 00:47:58.168897 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bhxt" event={"ID":"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846","Type":"ContainerStarted","Data":"986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98"} Jan 24 00:47:59 crc kubenswrapper[4676]: I0124 00:47:59.196033 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8bhxt" podStartSLOduration=3.621391472 podStartE2EDuration="9.196008987s" podCreationTimestamp="2026-01-24 00:47:50 +0000 UTC" firstStartedPulling="2026-01-24 00:47:52.119039181 +0000 UTC m=+2656.149010182" lastFinishedPulling="2026-01-24 00:47:57.693656696 +0000 UTC m=+2661.723627697" observedRunningTime="2026-01-24 00:47:59.193329194 +0000 UTC m=+2663.223300195" watchObservedRunningTime="2026-01-24 00:47:59.196008987 +0000 UTC m=+2663.225980008" Jan 24 00:48:00 crc kubenswrapper[4676]: I0124 00:48:00.774046 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:48:00 crc kubenswrapper[4676]: I0124 00:48:00.774091 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:48:01 crc kubenswrapper[4676]: I0124 00:48:01.822875 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8bhxt" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerName="registry-server" probeResult="failure" output=< Jan 24 00:48:01 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:48:01 crc kubenswrapper[4676]: > Jan 24 00:48:10 crc kubenswrapper[4676]: I0124 00:48:10.850149 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:48:10 crc kubenswrapper[4676]: I0124 00:48:10.924466 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:48:11 crc kubenswrapper[4676]: I0124 00:48:11.095489 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8bhxt"] Jan 24 00:48:12 crc kubenswrapper[4676]: I0124 00:48:12.315813 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8bhxt" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerName="registry-server" containerID="cri-o://986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98" gracePeriod=2 Jan 24 00:48:12 crc kubenswrapper[4676]: I0124 00:48:12.766112 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:48:12 crc kubenswrapper[4676]: I0124 00:48:12.887200 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-utilities\") pod \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " Jan 24 00:48:12 crc kubenswrapper[4676]: I0124 00:48:12.887298 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpkj4\" (UniqueName: \"kubernetes.io/projected/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-kube-api-access-rpkj4\") pod \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " Jan 24 00:48:12 crc kubenswrapper[4676]: I0124 00:48:12.887398 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-catalog-content\") pod \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\" (UID: \"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846\") " Jan 24 00:48:12 crc kubenswrapper[4676]: I0124 00:48:12.890114 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-utilities" (OuterVolumeSpecName: "utilities") pod "2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" (UID: "2f8aa49c-c56e-48e0-a8e2-e7a8777d4846"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:48:12 crc kubenswrapper[4676]: I0124 00:48:12.897993 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-kube-api-access-rpkj4" (OuterVolumeSpecName: "kube-api-access-rpkj4") pod "2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" (UID: "2f8aa49c-c56e-48e0-a8e2-e7a8777d4846"). InnerVolumeSpecName "kube-api-access-rpkj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:48:12 crc kubenswrapper[4676]: I0124 00:48:12.989768 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:12 crc kubenswrapper[4676]: I0124 00:48:12.989805 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpkj4\" (UniqueName: \"kubernetes.io/projected/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-kube-api-access-rpkj4\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.019026 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" (UID: "2f8aa49c-c56e-48e0-a8e2-e7a8777d4846"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.091798 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.331250 4676 generic.go:334] "Generic (PLEG): container finished" podID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerID="986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98" exitCode=0 Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.331340 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bhxt" event={"ID":"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846","Type":"ContainerDied","Data":"986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98"} Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.331435 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8bhxt" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.332082 4676 scope.go:117] "RemoveContainer" containerID="986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.331903 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8bhxt" event={"ID":"2f8aa49c-c56e-48e0-a8e2-e7a8777d4846","Type":"ContainerDied","Data":"b2d4b4da885e3e4644df949bcafb64ca58f900ed27867e562ecef9983b65ab9e"} Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.386759 4676 scope.go:117] "RemoveContainer" containerID="2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.409110 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8bhxt"] Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.423106 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8bhxt"] Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.424898 4676 scope.go:117] "RemoveContainer" containerID="b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.483134 4676 scope.go:117] "RemoveContainer" containerID="986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98" Jan 24 00:48:13 crc kubenswrapper[4676]: E0124 00:48:13.484095 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98\": container with ID starting with 986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98 not found: ID does not exist" containerID="986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.484157 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98"} err="failed to get container status \"986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98\": rpc error: code = NotFound desc = could not find container \"986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98\": container with ID starting with 986156910a9642cbf653dabbf5578f7a1410cb78bfa71ced2b3ac3f6b7af7a98 not found: ID does not exist" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.484195 4676 scope.go:117] "RemoveContainer" containerID="2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7" Jan 24 00:48:13 crc kubenswrapper[4676]: E0124 00:48:13.485464 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7\": container with ID starting with 2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7 not found: ID does not exist" containerID="2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.485697 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7"} err="failed to get container status \"2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7\": rpc error: code = NotFound desc = could not find container \"2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7\": container with ID starting with 2b74f2bb198b1e49d3579f7e148aa0f1446f32a9a24a52cd446f297186fa53a7 not found: ID does not exist" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.485907 4676 scope.go:117] "RemoveContainer" containerID="b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08" Jan 24 00:48:13 crc kubenswrapper[4676]: E0124 00:48:13.486595 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08\": container with ID starting with b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08 not found: ID does not exist" containerID="b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08" Jan 24 00:48:13 crc kubenswrapper[4676]: I0124 00:48:13.486658 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08"} err="failed to get container status \"b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08\": rpc error: code = NotFound desc = could not find container \"b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08\": container with ID starting with b5edc96625c9acbc57cac8015a5d029ceee5b34c3889182c899b340e2f68ab08 not found: ID does not exist" Jan 24 00:48:14 crc kubenswrapper[4676]: I0124 00:48:14.269281 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" path="/var/lib/kubelet/pods/2f8aa49c-c56e-48e0-a8e2-e7a8777d4846/volumes" Jan 24 00:48:33 crc kubenswrapper[4676]: I0124 00:48:33.547582 4676 generic.go:334] "Generic (PLEG): container finished" podID="75cae94d-b818-4f92-b42d-fd8cec63a657" containerID="9049acca3eaa19729de3abedb765017e8466c178a888c60c1ddeb96f4f0c0556" exitCode=0 Jan 24 00:48:33 crc kubenswrapper[4676]: I0124 00:48:33.547813 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" event={"ID":"75cae94d-b818-4f92-b42d-fd8cec63a657","Type":"ContainerDied","Data":"9049acca3eaa19729de3abedb765017e8466c178a888c60c1ddeb96f4f0c0556"} Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.016335 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.192006 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-combined-ca-bundle\") pod \"75cae94d-b818-4f92-b42d-fd8cec63a657\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.192280 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-0\") pod \"75cae94d-b818-4f92-b42d-fd8cec63a657\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.192443 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-ssh-key-openstack-edpm-ipam\") pod \"75cae94d-b818-4f92-b42d-fd8cec63a657\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.192588 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-1\") pod \"75cae94d-b818-4f92-b42d-fd8cec63a657\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.192681 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-1\") pod \"75cae94d-b818-4f92-b42d-fd8cec63a657\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.192771 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-inventory\") pod \"75cae94d-b818-4f92-b42d-fd8cec63a657\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.192852 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6689\" (UniqueName: \"kubernetes.io/projected/75cae94d-b818-4f92-b42d-fd8cec63a657-kube-api-access-g6689\") pod \"75cae94d-b818-4f92-b42d-fd8cec63a657\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.193001 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-extra-config-0\") pod \"75cae94d-b818-4f92-b42d-fd8cec63a657\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.193124 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-0\") pod \"75cae94d-b818-4f92-b42d-fd8cec63a657\" (UID: \"75cae94d-b818-4f92-b42d-fd8cec63a657\") " Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.213065 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75cae94d-b818-4f92-b42d-fd8cec63a657-kube-api-access-g6689" (OuterVolumeSpecName: "kube-api-access-g6689") pod "75cae94d-b818-4f92-b42d-fd8cec63a657" (UID: "75cae94d-b818-4f92-b42d-fd8cec63a657"). InnerVolumeSpecName "kube-api-access-g6689". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.213212 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "75cae94d-b818-4f92-b42d-fd8cec63a657" (UID: "75cae94d-b818-4f92-b42d-fd8cec63a657"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.227343 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "75cae94d-b818-4f92-b42d-fd8cec63a657" (UID: "75cae94d-b818-4f92-b42d-fd8cec63a657"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.231756 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "75cae94d-b818-4f92-b42d-fd8cec63a657" (UID: "75cae94d-b818-4f92-b42d-fd8cec63a657"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.232121 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "75cae94d-b818-4f92-b42d-fd8cec63a657" (UID: "75cae94d-b818-4f92-b42d-fd8cec63a657"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.233936 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "75cae94d-b818-4f92-b42d-fd8cec63a657" (UID: "75cae94d-b818-4f92-b42d-fd8cec63a657"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.254518 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "75cae94d-b818-4f92-b42d-fd8cec63a657" (UID: "75cae94d-b818-4f92-b42d-fd8cec63a657"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.262770 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-inventory" (OuterVolumeSpecName: "inventory") pod "75cae94d-b818-4f92-b42d-fd8cec63a657" (UID: "75cae94d-b818-4f92-b42d-fd8cec63a657"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.270116 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "75cae94d-b818-4f92-b42d-fd8cec63a657" (UID: "75cae94d-b818-4f92-b42d-fd8cec63a657"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.296425 4676 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.296455 4676 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.296464 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.296472 4676 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.296481 4676 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.296490 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.296498 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6689\" (UniqueName: \"kubernetes.io/projected/75cae94d-b818-4f92-b42d-fd8cec63a657-kube-api-access-g6689\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.296508 4676 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.296516 4676 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/75cae94d-b818-4f92-b42d-fd8cec63a657-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.562395 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" event={"ID":"75cae94d-b818-4f92-b42d-fd8cec63a657","Type":"ContainerDied","Data":"7a1a57426601586865e4d5bc7b431f54c628df289eb8caca5bd5436e21faee21"} Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.562647 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a1a57426601586865e4d5bc7b431f54c628df289eb8caca5bd5436e21faee21" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.562429 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-5cl2g" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.753804 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv"] Jan 24 00:48:35 crc kubenswrapper[4676]: E0124 00:48:35.754258 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerName="registry-server" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.754280 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerName="registry-server" Jan 24 00:48:35 crc kubenswrapper[4676]: E0124 00:48:35.754321 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerName="extract-content" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.754330 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerName="extract-content" Jan 24 00:48:35 crc kubenswrapper[4676]: E0124 00:48:35.754353 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75cae94d-b818-4f92-b42d-fd8cec63a657" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.754362 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="75cae94d-b818-4f92-b42d-fd8cec63a657" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 24 00:48:35 crc kubenswrapper[4676]: E0124 00:48:35.754397 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerName="extract-utilities" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.754406 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerName="extract-utilities" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.754641 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="75cae94d-b818-4f92-b42d-fd8cec63a657" containerName="nova-edpm-deployment-openstack-edpm-ipam" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.754671 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f8aa49c-c56e-48e0-a8e2-e7a8777d4846" containerName="registry-server" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.755668 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.758149 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.758158 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.758162 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.758449 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5vl7p" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.758616 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.765876 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv"] Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.909815 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.910089 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.910181 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcmjf\" (UniqueName: \"kubernetes.io/projected/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-kube-api-access-bcmjf\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.910426 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.910531 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.910613 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:35 crc kubenswrapper[4676]: I0124 00:48:35.910927 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.012349 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.012620 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.013060 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcmjf\" (UniqueName: \"kubernetes.io/projected/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-kube-api-access-bcmjf\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.013195 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.013273 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.013353 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.013553 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.017746 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.017958 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.018586 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.018831 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.018923 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.020288 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.032767 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcmjf\" (UniqueName: \"kubernetes.io/projected/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-kube-api-access-bcmjf\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.073520 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:48:36 crc kubenswrapper[4676]: I0124 00:48:36.620832 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv"] Jan 24 00:48:37 crc kubenswrapper[4676]: I0124 00:48:37.241413 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Jan 24 00:48:37 crc kubenswrapper[4676]: I0124 00:48:37.579049 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" event={"ID":"7ef31551-e4ed-48d0-a4d6-f9c2fb515966","Type":"ContainerStarted","Data":"49215d2f6ff846e5d4f5385e57b7216b7767e35f242d154ba2d3618223caeae2"} Jan 24 00:48:37 crc kubenswrapper[4676]: I0124 00:48:37.579350 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" event={"ID":"7ef31551-e4ed-48d0-a4d6-f9c2fb515966","Type":"ContainerStarted","Data":"8044fbd530591f1f5965e015b66ccd01325bde8aa28d26f8d37467c9160466d7"} Jan 24 00:48:37 crc kubenswrapper[4676]: I0124 00:48:37.601296 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" podStartSLOduration=1.985670659 podStartE2EDuration="2.601279921s" podCreationTimestamp="2026-01-24 00:48:35 +0000 UTC" firstStartedPulling="2026-01-24 00:48:36.621587448 +0000 UTC m=+2700.651558469" lastFinishedPulling="2026-01-24 00:48:37.23719669 +0000 UTC m=+2701.267167731" observedRunningTime="2026-01-24 00:48:37.598397722 +0000 UTC m=+2701.628368723" watchObservedRunningTime="2026-01-24 00:48:37.601279921 +0000 UTC m=+2701.631250922" Jan 24 00:49:09 crc kubenswrapper[4676]: I0124 00:49:09.364327 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:49:09 crc kubenswrapper[4676]: I0124 00:49:09.364908 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:49:39 crc kubenswrapper[4676]: I0124 00:49:39.364314 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:49:39 crc kubenswrapper[4676]: I0124 00:49:39.364820 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:50:09 crc kubenswrapper[4676]: I0124 00:50:09.364945 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:50:09 crc kubenswrapper[4676]: I0124 00:50:09.365796 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:50:09 crc kubenswrapper[4676]: I0124 00:50:09.365866 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:50:09 crc kubenswrapper[4676]: I0124 00:50:09.367256 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50556dab9c423476f3e768a6107a078181f27e372f40fe19e80cac55ab001315"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:50:09 crc kubenswrapper[4676]: I0124 00:50:09.367344 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://50556dab9c423476f3e768a6107a078181f27e372f40fe19e80cac55ab001315" gracePeriod=600 Jan 24 00:50:10 crc kubenswrapper[4676]: I0124 00:50:10.489193 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="50556dab9c423476f3e768a6107a078181f27e372f40fe19e80cac55ab001315" exitCode=0 Jan 24 00:50:10 crc kubenswrapper[4676]: I0124 00:50:10.489487 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"50556dab9c423476f3e768a6107a078181f27e372f40fe19e80cac55ab001315"} Jan 24 00:50:10 crc kubenswrapper[4676]: I0124 00:50:10.489937 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398"} Jan 24 00:50:10 crc kubenswrapper[4676]: I0124 00:50:10.489961 4676 scope.go:117] "RemoveContainer" containerID="0cc55a19ee9f58fa576e51c2d6dcb1203ed54bdb81c4ec86abf6a39092861cac" Jan 24 00:51:59 crc kubenswrapper[4676]: I0124 00:51:59.640627 4676 generic.go:334] "Generic (PLEG): container finished" podID="7ef31551-e4ed-48d0-a4d6-f9c2fb515966" containerID="49215d2f6ff846e5d4f5385e57b7216b7767e35f242d154ba2d3618223caeae2" exitCode=0 Jan 24 00:51:59 crc kubenswrapper[4676]: I0124 00:51:59.640836 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" event={"ID":"7ef31551-e4ed-48d0-a4d6-f9c2fb515966","Type":"ContainerDied","Data":"49215d2f6ff846e5d4f5385e57b7216b7767e35f242d154ba2d3618223caeae2"} Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.127670 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.268698 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-inventory\") pod \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.269089 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-2\") pod \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.269136 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-0\") pod \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.269236 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-telemetry-combined-ca-bundle\") pod \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.269363 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-1\") pod \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.269458 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ssh-key-openstack-edpm-ipam\") pod \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.270492 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcmjf\" (UniqueName: \"kubernetes.io/projected/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-kube-api-access-bcmjf\") pod \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\" (UID: \"7ef31551-e4ed-48d0-a4d6-f9c2fb515966\") " Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.277112 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-kube-api-access-bcmjf" (OuterVolumeSpecName: "kube-api-access-bcmjf") pod "7ef31551-e4ed-48d0-a4d6-f9c2fb515966" (UID: "7ef31551-e4ed-48d0-a4d6-f9c2fb515966"). InnerVolumeSpecName "kube-api-access-bcmjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.277504 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7ef31551-e4ed-48d0-a4d6-f9c2fb515966" (UID: "7ef31551-e4ed-48d0-a4d6-f9c2fb515966"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.298355 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7ef31551-e4ed-48d0-a4d6-f9c2fb515966" (UID: "7ef31551-e4ed-48d0-a4d6-f9c2fb515966"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.303851 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-inventory" (OuterVolumeSpecName: "inventory") pod "7ef31551-e4ed-48d0-a4d6-f9c2fb515966" (UID: "7ef31551-e4ed-48d0-a4d6-f9c2fb515966"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.308648 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "7ef31551-e4ed-48d0-a4d6-f9c2fb515966" (UID: "7ef31551-e4ed-48d0-a4d6-f9c2fb515966"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.312554 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "7ef31551-e4ed-48d0-a4d6-f9c2fb515966" (UID: "7ef31551-e4ed-48d0-a4d6-f9c2fb515966"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.327807 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "7ef31551-e4ed-48d0-a4d6-f9c2fb515966" (UID: "7ef31551-e4ed-48d0-a4d6-f9c2fb515966"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.373129 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.373169 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcmjf\" (UniqueName: \"kubernetes.io/projected/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-kube-api-access-bcmjf\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.373183 4676 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-inventory\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.373195 4676 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.373207 4676 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.373223 4676 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.373236 4676 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7ef31551-e4ed-48d0-a4d6-f9c2fb515966-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.663786 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" event={"ID":"7ef31551-e4ed-48d0-a4d6-f9c2fb515966","Type":"ContainerDied","Data":"8044fbd530591f1f5965e015b66ccd01325bde8aa28d26f8d37467c9160466d7"} Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.663830 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8044fbd530591f1f5965e015b66ccd01325bde8aa28d26f8d37467c9160466d7" Jan 24 00:52:01 crc kubenswrapper[4676]: I0124 00:52:01.663884 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv" Jan 24 00:52:09 crc kubenswrapper[4676]: I0124 00:52:09.364836 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:52:09 crc kubenswrapper[4676]: I0124 00:52:09.365587 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.142848 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4hpk9"] Jan 24 00:52:26 crc kubenswrapper[4676]: E0124 00:52:26.144875 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ef31551-e4ed-48d0-a4d6-f9c2fb515966" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.144994 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ef31551-e4ed-48d0-a4d6-f9c2fb515966" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.145284 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ef31551-e4ed-48d0-a4d6-f9c2fb515966" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.147289 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.168197 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hpk9"] Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.301750 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-utilities\") pod \"redhat-marketplace-4hpk9\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.301936 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-catalog-content\") pod \"redhat-marketplace-4hpk9\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.301965 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct2bk\" (UniqueName: \"kubernetes.io/projected/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-kube-api-access-ct2bk\") pod \"redhat-marketplace-4hpk9\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.403017 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-catalog-content\") pod \"redhat-marketplace-4hpk9\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.403068 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct2bk\" (UniqueName: \"kubernetes.io/projected/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-kube-api-access-ct2bk\") pod \"redhat-marketplace-4hpk9\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.403156 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-utilities\") pod \"redhat-marketplace-4hpk9\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.404921 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-catalog-content\") pod \"redhat-marketplace-4hpk9\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.404942 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-utilities\") pod \"redhat-marketplace-4hpk9\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.425357 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct2bk\" (UniqueName: \"kubernetes.io/projected/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-kube-api-access-ct2bk\") pod \"redhat-marketplace-4hpk9\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.465867 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:26 crc kubenswrapper[4676]: I0124 00:52:26.981658 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hpk9"] Jan 24 00:52:26 crc kubenswrapper[4676]: W0124 00:52:26.987616 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod521707d5_eb8e_4c5c_b171_2a001bcdf0a1.slice/crio-b819727379ac8360ec1c277c305e34faeb524ceba12d9ad37312d65500a3c018 WatchSource:0}: Error finding container b819727379ac8360ec1c277c305e34faeb524ceba12d9ad37312d65500a3c018: Status 404 returned error can't find the container with id b819727379ac8360ec1c277c305e34faeb524ceba12d9ad37312d65500a3c018 Jan 24 00:52:27 crc kubenswrapper[4676]: I0124 00:52:27.921433 4676 generic.go:334] "Generic (PLEG): container finished" podID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerID="3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b" exitCode=0 Jan 24 00:52:27 crc kubenswrapper[4676]: I0124 00:52:27.921573 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hpk9" event={"ID":"521707d5-eb8e-4c5c-b171-2a001bcdf0a1","Type":"ContainerDied","Data":"3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b"} Jan 24 00:52:27 crc kubenswrapper[4676]: I0124 00:52:27.926493 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hpk9" event={"ID":"521707d5-eb8e-4c5c-b171-2a001bcdf0a1","Type":"ContainerStarted","Data":"b819727379ac8360ec1c277c305e34faeb524ceba12d9ad37312d65500a3c018"} Jan 24 00:52:27 crc kubenswrapper[4676]: I0124 00:52:27.926552 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:52:28 crc kubenswrapper[4676]: I0124 00:52:28.936241 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hpk9" event={"ID":"521707d5-eb8e-4c5c-b171-2a001bcdf0a1","Type":"ContainerStarted","Data":"88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c"} Jan 24 00:52:29 crc kubenswrapper[4676]: I0124 00:52:29.951425 4676 generic.go:334] "Generic (PLEG): container finished" podID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerID="88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c" exitCode=0 Jan 24 00:52:29 crc kubenswrapper[4676]: I0124 00:52:29.951486 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hpk9" event={"ID":"521707d5-eb8e-4c5c-b171-2a001bcdf0a1","Type":"ContainerDied","Data":"88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c"} Jan 24 00:52:30 crc kubenswrapper[4676]: I0124 00:52:30.967495 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hpk9" event={"ID":"521707d5-eb8e-4c5c-b171-2a001bcdf0a1","Type":"ContainerStarted","Data":"e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d"} Jan 24 00:52:30 crc kubenswrapper[4676]: I0124 00:52:30.996624 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4hpk9" podStartSLOduration=2.433693255 podStartE2EDuration="4.996609787s" podCreationTimestamp="2026-01-24 00:52:26 +0000 UTC" firstStartedPulling="2026-01-24 00:52:27.924283229 +0000 UTC m=+2931.954254270" lastFinishedPulling="2026-01-24 00:52:30.487199801 +0000 UTC m=+2934.517170802" observedRunningTime="2026-01-24 00:52:30.99146194 +0000 UTC m=+2935.021432971" watchObservedRunningTime="2026-01-24 00:52:30.996609787 +0000 UTC m=+2935.026580778" Jan 24 00:52:36 crc kubenswrapper[4676]: I0124 00:52:36.466459 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:36 crc kubenswrapper[4676]: I0124 00:52:36.467101 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:36 crc kubenswrapper[4676]: I0124 00:52:36.539248 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:37 crc kubenswrapper[4676]: I0124 00:52:37.125726 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:37 crc kubenswrapper[4676]: I0124 00:52:37.191316 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hpk9"] Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.069312 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4hpk9" podUID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerName="registry-server" containerID="cri-o://e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d" gracePeriod=2 Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.364952 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.365011 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.595733 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.720302 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct2bk\" (UniqueName: \"kubernetes.io/projected/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-kube-api-access-ct2bk\") pod \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.720585 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-utilities\") pod \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.720660 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-catalog-content\") pod \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\" (UID: \"521707d5-eb8e-4c5c-b171-2a001bcdf0a1\") " Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.721357 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-utilities" (OuterVolumeSpecName: "utilities") pod "521707d5-eb8e-4c5c-b171-2a001bcdf0a1" (UID: "521707d5-eb8e-4c5c-b171-2a001bcdf0a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.729253 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-kube-api-access-ct2bk" (OuterVolumeSpecName: "kube-api-access-ct2bk") pod "521707d5-eb8e-4c5c-b171-2a001bcdf0a1" (UID: "521707d5-eb8e-4c5c-b171-2a001bcdf0a1"). InnerVolumeSpecName "kube-api-access-ct2bk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.764402 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "521707d5-eb8e-4c5c-b171-2a001bcdf0a1" (UID: "521707d5-eb8e-4c5c-b171-2a001bcdf0a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.823978 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.824285 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:39 crc kubenswrapper[4676]: I0124 00:52:39.824566 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct2bk\" (UniqueName: \"kubernetes.io/projected/521707d5-eb8e-4c5c-b171-2a001bcdf0a1-kube-api-access-ct2bk\") on node \"crc\" DevicePath \"\"" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.084088 4676 generic.go:334] "Generic (PLEG): container finished" podID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerID="e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d" exitCode=0 Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.084151 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hpk9" event={"ID":"521707d5-eb8e-4c5c-b171-2a001bcdf0a1","Type":"ContainerDied","Data":"e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d"} Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.084183 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4hpk9" event={"ID":"521707d5-eb8e-4c5c-b171-2a001bcdf0a1","Type":"ContainerDied","Data":"b819727379ac8360ec1c277c305e34faeb524ceba12d9ad37312d65500a3c018"} Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.084203 4676 scope.go:117] "RemoveContainer" containerID="e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.084460 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4hpk9" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.142183 4676 scope.go:117] "RemoveContainer" containerID="88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.151238 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hpk9"] Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.176018 4676 scope.go:117] "RemoveContainer" containerID="3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.179168 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4hpk9"] Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.231181 4676 scope.go:117] "RemoveContainer" containerID="e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d" Jan 24 00:52:40 crc kubenswrapper[4676]: E0124 00:52:40.232223 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d\": container with ID starting with e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d not found: ID does not exist" containerID="e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.232297 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d"} err="failed to get container status \"e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d\": rpc error: code = NotFound desc = could not find container \"e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d\": container with ID starting with e41843d9624da2bf74d0d9538e6a112fc9774350c9ef35f851f14169594e626d not found: ID does not exist" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.232345 4676 scope.go:117] "RemoveContainer" containerID="88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c" Jan 24 00:52:40 crc kubenswrapper[4676]: E0124 00:52:40.233326 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c\": container with ID starting with 88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c not found: ID does not exist" containerID="88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.233409 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c"} err="failed to get container status \"88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c\": rpc error: code = NotFound desc = could not find container \"88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c\": container with ID starting with 88d697e7ed40846c3bb522d739af626a9f6db8495cc8211d536e953fcc0b640c not found: ID does not exist" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.233462 4676 scope.go:117] "RemoveContainer" containerID="3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b" Jan 24 00:52:40 crc kubenswrapper[4676]: E0124 00:52:40.233947 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b\": container with ID starting with 3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b not found: ID does not exist" containerID="3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.234023 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b"} err="failed to get container status \"3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b\": rpc error: code = NotFound desc = could not find container \"3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b\": container with ID starting with 3bb1f921af66f44c77cccc2df88d800be23799a13d52bafb67d019e34d21813b not found: ID does not exist" Jan 24 00:52:40 crc kubenswrapper[4676]: I0124 00:52:40.274405 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" path="/var/lib/kubelet/pods/521707d5-eb8e-4c5c-b171-2a001bcdf0a1/volumes" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.927045 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Jan 24 00:52:59 crc kubenswrapper[4676]: E0124 00:52:59.928589 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerName="extract-content" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.928627 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerName="extract-content" Jan 24 00:52:59 crc kubenswrapper[4676]: E0124 00:52:59.928684 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerName="registry-server" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.928705 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerName="registry-server" Jan 24 00:52:59 crc kubenswrapper[4676]: E0124 00:52:59.928743 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerName="extract-utilities" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.928764 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerName="extract-utilities" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.929209 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="521707d5-eb8e-4c5c-b171-2a001bcdf0a1" containerName="registry-server" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.930511 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.935460 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.935535 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.936585 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.939032 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-v9f6d" Jan 24 00:52:59 crc kubenswrapper[4676]: I0124 00:52:59.947610 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.062305 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.062409 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.062502 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwz6w\" (UniqueName: \"kubernetes.io/projected/3e2adf44-9053-4dcd-9d47-27910710dbc8-kube-api-access-dwz6w\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.062663 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.062796 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-config-data\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.062860 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.062933 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.063008 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.063037 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.165092 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.165550 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-config-data\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.165761 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.165998 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.166214 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.166426 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.166688 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.165764 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.166909 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.167185 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.167225 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.167227 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwz6w\" (UniqueName: \"kubernetes.io/projected/3e2adf44-9053-4dcd-9d47-27910710dbc8-kube-api-access-dwz6w\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.168280 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.171213 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-config-data\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.178224 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.179271 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.179629 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.193112 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwz6w\" (UniqueName: \"kubernetes.io/projected/3e2adf44-9053-4dcd-9d47-27910710dbc8-kube-api-access-dwz6w\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.213692 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.264666 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 24 00:53:00 crc kubenswrapper[4676]: I0124 00:53:00.768190 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Jan 24 00:53:01 crc kubenswrapper[4676]: I0124 00:53:01.296320 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3e2adf44-9053-4dcd-9d47-27910710dbc8","Type":"ContainerStarted","Data":"7398b97beedabcbadf8e82b5f7009c22665a40a681f29f7b52fa09e6bbc8c295"} Jan 24 00:53:09 crc kubenswrapper[4676]: I0124 00:53:09.363768 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 00:53:09 crc kubenswrapper[4676]: I0124 00:53:09.364345 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 00:53:09 crc kubenswrapper[4676]: I0124 00:53:09.364386 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 00:53:09 crc kubenswrapper[4676]: I0124 00:53:09.365087 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 00:53:09 crc kubenswrapper[4676]: I0124 00:53:09.365130 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" gracePeriod=600 Jan 24 00:53:09 crc kubenswrapper[4676]: E0124 00:53:09.990882 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:53:10 crc kubenswrapper[4676]: I0124 00:53:10.383876 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" exitCode=0 Jan 24 00:53:10 crc kubenswrapper[4676]: I0124 00:53:10.383923 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398"} Jan 24 00:53:10 crc kubenswrapper[4676]: I0124 00:53:10.383954 4676 scope.go:117] "RemoveContainer" containerID="50556dab9c423476f3e768a6107a078181f27e372f40fe19e80cac55ab001315" Jan 24 00:53:10 crc kubenswrapper[4676]: I0124 00:53:10.384654 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:53:10 crc kubenswrapper[4676]: E0124 00:53:10.384881 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:53:22 crc kubenswrapper[4676]: I0124 00:53:22.255393 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:53:22 crc kubenswrapper[4676]: E0124 00:53:22.256170 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:53:33 crc kubenswrapper[4676]: E0124 00:53:33.674358 4676 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Jan 24 00:53:33 crc kubenswrapper[4676]: E0124 00:53:33.677316 4676 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dwz6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(3e2adf44-9053-4dcd-9d47-27910710dbc8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 24 00:53:33 crc kubenswrapper[4676]: E0124 00:53:33.678713 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="3e2adf44-9053-4dcd-9d47-27910710dbc8" Jan 24 00:53:33 crc kubenswrapper[4676]: E0124 00:53:33.707158 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="3e2adf44-9053-4dcd-9d47-27910710dbc8" Jan 24 00:53:35 crc kubenswrapper[4676]: I0124 00:53:35.256004 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:53:35 crc kubenswrapper[4676]: E0124 00:53:35.256680 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:53:47 crc kubenswrapper[4676]: I0124 00:53:47.730835 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Jan 24 00:53:48 crc kubenswrapper[4676]: I0124 00:53:48.841904 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3e2adf44-9053-4dcd-9d47-27910710dbc8","Type":"ContainerStarted","Data":"5ce215c83e117e67d9aa3269233be260f6fe19170cac686c5414d50aaa942b7b"} Jan 24 00:53:50 crc kubenswrapper[4676]: I0124 00:53:50.257541 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:53:50 crc kubenswrapper[4676]: E0124 00:53:50.259701 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.114651 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=17.155567276 podStartE2EDuration="1m4.114635683s" podCreationTimestamp="2026-01-24 00:52:58 +0000 UTC" firstStartedPulling="2026-01-24 00:53:00.769598887 +0000 UTC m=+2964.799569898" lastFinishedPulling="2026-01-24 00:53:47.728667304 +0000 UTC m=+3011.758638305" observedRunningTime="2026-01-24 00:53:48.865678683 +0000 UTC m=+3012.895649684" watchObservedRunningTime="2026-01-24 00:54:02.114635683 +0000 UTC m=+3026.144606684" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.123322 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vsfsj"] Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.125750 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.201553 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vsfsj"] Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.292506 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34862c1d-2d18-42f8-9ef7-71d349c019fd-catalog-content\") pod \"certified-operators-vsfsj\" (UID: \"34862c1d-2d18-42f8-9ef7-71d349c019fd\") " pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.292565 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvh2j\" (UniqueName: \"kubernetes.io/projected/34862c1d-2d18-42f8-9ef7-71d349c019fd-kube-api-access-cvh2j\") pod \"certified-operators-vsfsj\" (UID: \"34862c1d-2d18-42f8-9ef7-71d349c019fd\") " pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.292584 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34862c1d-2d18-42f8-9ef7-71d349c019fd-utilities\") pod \"certified-operators-vsfsj\" (UID: \"34862c1d-2d18-42f8-9ef7-71d349c019fd\") " pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.395161 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34862c1d-2d18-42f8-9ef7-71d349c019fd-catalog-content\") pod \"certified-operators-vsfsj\" (UID: \"34862c1d-2d18-42f8-9ef7-71d349c019fd\") " pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.395236 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvh2j\" (UniqueName: \"kubernetes.io/projected/34862c1d-2d18-42f8-9ef7-71d349c019fd-kube-api-access-cvh2j\") pod \"certified-operators-vsfsj\" (UID: \"34862c1d-2d18-42f8-9ef7-71d349c019fd\") " pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.395255 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34862c1d-2d18-42f8-9ef7-71d349c019fd-utilities\") pod \"certified-operators-vsfsj\" (UID: \"34862c1d-2d18-42f8-9ef7-71d349c019fd\") " pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.395792 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34862c1d-2d18-42f8-9ef7-71d349c019fd-catalog-content\") pod \"certified-operators-vsfsj\" (UID: \"34862c1d-2d18-42f8-9ef7-71d349c019fd\") " pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.395829 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34862c1d-2d18-42f8-9ef7-71d349c019fd-utilities\") pod \"certified-operators-vsfsj\" (UID: \"34862c1d-2d18-42f8-9ef7-71d349c019fd\") " pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.412300 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvh2j\" (UniqueName: \"kubernetes.io/projected/34862c1d-2d18-42f8-9ef7-71d349c019fd-kube-api-access-cvh2j\") pod \"certified-operators-vsfsj\" (UID: \"34862c1d-2d18-42f8-9ef7-71d349c019fd\") " pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:02 crc kubenswrapper[4676]: I0124 00:54:02.450884 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:03 crc kubenswrapper[4676]: I0124 00:54:03.089147 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vsfsj"] Jan 24 00:54:04 crc kubenswrapper[4676]: I0124 00:54:03.999395 4676 generic.go:334] "Generic (PLEG): container finished" podID="34862c1d-2d18-42f8-9ef7-71d349c019fd" containerID="bd0522c40e1ed2b9f90bb6ab56e9ed807a381792f0d150f9ebf635515f27885b" exitCode=0 Jan 24 00:54:04 crc kubenswrapper[4676]: I0124 00:54:03.999521 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsfsj" event={"ID":"34862c1d-2d18-42f8-9ef7-71d349c019fd","Type":"ContainerDied","Data":"bd0522c40e1ed2b9f90bb6ab56e9ed807a381792f0d150f9ebf635515f27885b"} Jan 24 00:54:04 crc kubenswrapper[4676]: I0124 00:54:03.999736 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsfsj" event={"ID":"34862c1d-2d18-42f8-9ef7-71d349c019fd","Type":"ContainerStarted","Data":"e6f60f5616d0008aaa8f773063b8843a5b35c44a564220184ce620f19abe2326"} Jan 24 00:54:05 crc kubenswrapper[4676]: I0124 00:54:05.256357 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:54:05 crc kubenswrapper[4676]: E0124 00:54:05.256873 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:54:10 crc kubenswrapper[4676]: I0124 00:54:10.051208 4676 generic.go:334] "Generic (PLEG): container finished" podID="34862c1d-2d18-42f8-9ef7-71d349c019fd" containerID="e9f6ee47231da28f152e04b892dfc525d731c8ebb15f49b6feabec250f0eb303" exitCode=0 Jan 24 00:54:10 crc kubenswrapper[4676]: I0124 00:54:10.051339 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsfsj" event={"ID":"34862c1d-2d18-42f8-9ef7-71d349c019fd","Type":"ContainerDied","Data":"e9f6ee47231da28f152e04b892dfc525d731c8ebb15f49b6feabec250f0eb303"} Jan 24 00:54:11 crc kubenswrapper[4676]: I0124 00:54:11.069173 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vsfsj" event={"ID":"34862c1d-2d18-42f8-9ef7-71d349c019fd","Type":"ContainerStarted","Data":"6f5bcd53abfa53e8152928e53d01c7a3fab65ae2fc2e34afd71588e99e2c7975"} Jan 24 00:54:11 crc kubenswrapper[4676]: I0124 00:54:11.101580 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vsfsj" podStartSLOduration=2.6596724849999998 podStartE2EDuration="9.101561149s" podCreationTimestamp="2026-01-24 00:54:02 +0000 UTC" firstStartedPulling="2026-01-24 00:54:04.002138884 +0000 UTC m=+3028.032109885" lastFinishedPulling="2026-01-24 00:54:10.444027548 +0000 UTC m=+3034.473998549" observedRunningTime="2026-01-24 00:54:11.091024685 +0000 UTC m=+3035.120995686" watchObservedRunningTime="2026-01-24 00:54:11.101561149 +0000 UTC m=+3035.131532150" Jan 24 00:54:12 crc kubenswrapper[4676]: I0124 00:54:12.725788 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:12 crc kubenswrapper[4676]: I0124 00:54:12.728219 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:12 crc kubenswrapper[4676]: I0124 00:54:12.812953 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:19 crc kubenswrapper[4676]: I0124 00:54:19.256292 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:54:19 crc kubenswrapper[4676]: E0124 00:54:19.257113 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:54:22 crc kubenswrapper[4676]: I0124 00:54:22.502171 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vsfsj" Jan 24 00:54:22 crc kubenswrapper[4676]: I0124 00:54:22.587548 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vsfsj"] Jan 24 00:54:22 crc kubenswrapper[4676]: I0124 00:54:22.650364 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h6lxx"] Jan 24 00:54:22 crc kubenswrapper[4676]: I0124 00:54:22.650662 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-h6lxx" podUID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerName="registry-server" containerID="cri-o://08fb9c82277ed77127a24cba780debc02f8a59da1333b461232dac0199f577d3" gracePeriod=2 Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.188495 4676 generic.go:334] "Generic (PLEG): container finished" podID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerID="08fb9c82277ed77127a24cba780debc02f8a59da1333b461232dac0199f577d3" exitCode=0 Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.188738 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6lxx" event={"ID":"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec","Type":"ContainerDied","Data":"08fb9c82277ed77127a24cba780debc02f8a59da1333b461232dac0199f577d3"} Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.189083 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-h6lxx" event={"ID":"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec","Type":"ContainerDied","Data":"fb5783295da5666c028415e55625575a6eddcc0d1dfa67d4345fe77ee0fb6e6b"} Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.189115 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb5783295da5666c028415e55625575a6eddcc0d1dfa67d4345fe77ee0fb6e6b" Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.227325 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.254084 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdlrw\" (UniqueName: \"kubernetes.io/projected/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-kube-api-access-wdlrw\") pod \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.254156 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-utilities\") pod \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.254186 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-catalog-content\") pod \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\" (UID: \"8b3f0cd4-dad6-4882-8cf8-03e0a23768ec\") " Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.255069 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-utilities" (OuterVolumeSpecName: "utilities") pod "8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" (UID: "8b3f0cd4-dad6-4882-8cf8-03e0a23768ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.262719 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-kube-api-access-wdlrw" (OuterVolumeSpecName: "kube-api-access-wdlrw") pod "8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" (UID: "8b3f0cd4-dad6-4882-8cf8-03e0a23768ec"). InnerVolumeSpecName "kube-api-access-wdlrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.315647 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" (UID: "8b3f0cd4-dad6-4882-8cf8-03e0a23768ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.356126 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdlrw\" (UniqueName: \"kubernetes.io/projected/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-kube-api-access-wdlrw\") on node \"crc\" DevicePath \"\"" Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.356162 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:54:23 crc kubenswrapper[4676]: I0124 00:54:23.356171 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:54:24 crc kubenswrapper[4676]: I0124 00:54:24.195525 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-h6lxx" Jan 24 00:54:24 crc kubenswrapper[4676]: I0124 00:54:24.223564 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-h6lxx"] Jan 24 00:54:24 crc kubenswrapper[4676]: I0124 00:54:24.235411 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-h6lxx"] Jan 24 00:54:24 crc kubenswrapper[4676]: I0124 00:54:24.267895 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" path="/var/lib/kubelet/pods/8b3f0cd4-dad6-4882-8cf8-03e0a23768ec/volumes" Jan 24 00:54:32 crc kubenswrapper[4676]: I0124 00:54:32.256334 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:54:32 crc kubenswrapper[4676]: E0124 00:54:32.256996 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:54:42 crc kubenswrapper[4676]: I0124 00:54:42.970200 4676 scope.go:117] "RemoveContainer" containerID="e1126bba560d6ad086386fdbfde6ce098ca6ec45eb8ef89aa5c2a014e09ba3c2" Jan 24 00:54:43 crc kubenswrapper[4676]: I0124 00:54:43.002556 4676 scope.go:117] "RemoveContainer" containerID="08fb9c82277ed77127a24cba780debc02f8a59da1333b461232dac0199f577d3" Jan 24 00:54:43 crc kubenswrapper[4676]: I0124 00:54:43.035683 4676 scope.go:117] "RemoveContainer" containerID="52bc77f8ab557f140d976553f763b957c850be181430d480625497a667852cf5" Jan 24 00:54:47 crc kubenswrapper[4676]: I0124 00:54:47.256498 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:54:47 crc kubenswrapper[4676]: E0124 00:54:47.257442 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:55:01 crc kubenswrapper[4676]: I0124 00:55:01.256669 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:55:01 crc kubenswrapper[4676]: E0124 00:55:01.257663 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:55:15 crc kubenswrapper[4676]: I0124 00:55:15.256589 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:55:15 crc kubenswrapper[4676]: E0124 00:55:15.259844 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:55:15 crc kubenswrapper[4676]: I0124 00:55:15.844036 4676 generic.go:334] "Generic (PLEG): container finished" podID="3e2adf44-9053-4dcd-9d47-27910710dbc8" containerID="5ce215c83e117e67d9aa3269233be260f6fe19170cac686c5414d50aaa942b7b" exitCode=0 Jan 24 00:55:15 crc kubenswrapper[4676]: I0124 00:55:15.844099 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3e2adf44-9053-4dcd-9d47-27910710dbc8","Type":"ContainerDied","Data":"5ce215c83e117e67d9aa3269233be260f6fe19170cac686c5414d50aaa942b7b"} Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.461598 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.572534 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config-secret\") pod \"3e2adf44-9053-4dcd-9d47-27910710dbc8\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.572599 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ca-certs\") pod \"3e2adf44-9053-4dcd-9d47-27910710dbc8\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.572671 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ssh-key\") pod \"3e2adf44-9053-4dcd-9d47-27910710dbc8\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.572729 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-config-data\") pod \"3e2adf44-9053-4dcd-9d47-27910710dbc8\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.572798 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-workdir\") pod \"3e2adf44-9053-4dcd-9d47-27910710dbc8\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.572821 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"3e2adf44-9053-4dcd-9d47-27910710dbc8\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.572892 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config\") pod \"3e2adf44-9053-4dcd-9d47-27910710dbc8\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.572970 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-temporary\") pod \"3e2adf44-9053-4dcd-9d47-27910710dbc8\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.573015 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwz6w\" (UniqueName: \"kubernetes.io/projected/3e2adf44-9053-4dcd-9d47-27910710dbc8-kube-api-access-dwz6w\") pod \"3e2adf44-9053-4dcd-9d47-27910710dbc8\" (UID: \"3e2adf44-9053-4dcd-9d47-27910710dbc8\") " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.575282 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-config-data" (OuterVolumeSpecName: "config-data") pod "3e2adf44-9053-4dcd-9d47-27910710dbc8" (UID: "3e2adf44-9053-4dcd-9d47-27910710dbc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.576154 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "3e2adf44-9053-4dcd-9d47-27910710dbc8" (UID: "3e2adf44-9053-4dcd-9d47-27910710dbc8"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.580888 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "3e2adf44-9053-4dcd-9d47-27910710dbc8" (UID: "3e2adf44-9053-4dcd-9d47-27910710dbc8"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.585878 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "3e2adf44-9053-4dcd-9d47-27910710dbc8" (UID: "3e2adf44-9053-4dcd-9d47-27910710dbc8"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.602574 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e2adf44-9053-4dcd-9d47-27910710dbc8-kube-api-access-dwz6w" (OuterVolumeSpecName: "kube-api-access-dwz6w") pod "3e2adf44-9053-4dcd-9d47-27910710dbc8" (UID: "3e2adf44-9053-4dcd-9d47-27910710dbc8"). InnerVolumeSpecName "kube-api-access-dwz6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.608583 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "3e2adf44-9053-4dcd-9d47-27910710dbc8" (UID: "3e2adf44-9053-4dcd-9d47-27910710dbc8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.621810 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3e2adf44-9053-4dcd-9d47-27910710dbc8" (UID: "3e2adf44-9053-4dcd-9d47-27910710dbc8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.623696 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "3e2adf44-9053-4dcd-9d47-27910710dbc8" (UID: "3e2adf44-9053-4dcd-9d47-27910710dbc8"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.642933 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "3e2adf44-9053-4dcd-9d47-27910710dbc8" (UID: "3e2adf44-9053-4dcd-9d47-27910710dbc8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.674954 4676 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.675035 4676 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.675060 4676 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config\") on node \"crc\" DevicePath \"\"" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.675083 4676 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3e2adf44-9053-4dcd-9d47-27910710dbc8-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.675102 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwz6w\" (UniqueName: \"kubernetes.io/projected/3e2adf44-9053-4dcd-9d47-27910710dbc8-kube-api-access-dwz6w\") on node \"crc\" DevicePath \"\"" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.675121 4676 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.675139 4676 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ca-certs\") on node \"crc\" DevicePath \"\"" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.675177 4676 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3e2adf44-9053-4dcd-9d47-27910710dbc8-ssh-key\") on node \"crc\" DevicePath \"\"" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.675193 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e2adf44-9053-4dcd-9d47-27910710dbc8-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.705941 4676 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.781111 4676 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.866913 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3e2adf44-9053-4dcd-9d47-27910710dbc8","Type":"ContainerDied","Data":"7398b97beedabcbadf8e82b5f7009c22665a40a681f29f7b52fa09e6bbc8c295"} Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.866972 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7398b97beedabcbadf8e82b5f7009c22665a40a681f29f7b52fa09e6bbc8c295" Jan 24 00:55:17 crc kubenswrapper[4676]: I0124 00:55:17.867073 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.125161 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 24 00:55:21 crc kubenswrapper[4676]: E0124 00:55:21.126605 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerName="extract-utilities" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.126637 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerName="extract-utilities" Jan 24 00:55:21 crc kubenswrapper[4676]: E0124 00:55:21.126671 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e2adf44-9053-4dcd-9d47-27910710dbc8" containerName="tempest-tests-tempest-tests-runner" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.126686 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e2adf44-9053-4dcd-9d47-27910710dbc8" containerName="tempest-tests-tempest-tests-runner" Jan 24 00:55:21 crc kubenswrapper[4676]: E0124 00:55:21.126714 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerName="registry-server" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.126728 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerName="registry-server" Jan 24 00:55:21 crc kubenswrapper[4676]: E0124 00:55:21.126778 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerName="extract-content" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.126792 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerName="extract-content" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.127176 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e2adf44-9053-4dcd-9d47-27910710dbc8" containerName="tempest-tests-tempest-tests-runner" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.127212 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b3f0cd4-dad6-4882-8cf8-03e0a23768ec" containerName="registry-server" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.128369 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.132437 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-v9f6d" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.140053 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.253105 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dsvg\" (UniqueName: \"kubernetes.io/projected/e1574f42-e89a-40d4-b6da-2d4ef0824916-kube-api-access-2dsvg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e1574f42-e89a-40d4-b6da-2d4ef0824916\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.253341 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e1574f42-e89a-40d4-b6da-2d4ef0824916\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.357447 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dsvg\" (UniqueName: \"kubernetes.io/projected/e1574f42-e89a-40d4-b6da-2d4ef0824916-kube-api-access-2dsvg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e1574f42-e89a-40d4-b6da-2d4ef0824916\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.357586 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e1574f42-e89a-40d4-b6da-2d4ef0824916\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.358725 4676 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e1574f42-e89a-40d4-b6da-2d4ef0824916\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.408879 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dsvg\" (UniqueName: \"kubernetes.io/projected/e1574f42-e89a-40d4-b6da-2d4ef0824916-kube-api-access-2dsvg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e1574f42-e89a-40d4-b6da-2d4ef0824916\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.410423 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"e1574f42-e89a-40d4-b6da-2d4ef0824916\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.462511 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Jan 24 00:55:21 crc kubenswrapper[4676]: I0124 00:55:21.937812 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Jan 24 00:55:22 crc kubenswrapper[4676]: I0124 00:55:22.924751 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e1574f42-e89a-40d4-b6da-2d4ef0824916","Type":"ContainerStarted","Data":"0fe4334d32c0067c373d345a2e81177a0799a21f58306691839b8dcf3179e435"} Jan 24 00:55:23 crc kubenswrapper[4676]: I0124 00:55:23.939706 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"e1574f42-e89a-40d4-b6da-2d4ef0824916","Type":"ContainerStarted","Data":"57f025dc489d997bdebb6abcb40b61a62627feda286f2eea27f4307b37064031"} Jan 24 00:55:23 crc kubenswrapper[4676]: I0124 00:55:23.963491 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.968105555 podStartE2EDuration="2.963461929s" podCreationTimestamp="2026-01-24 00:55:21 +0000 UTC" firstStartedPulling="2026-01-24 00:55:21.943934167 +0000 UTC m=+3105.973905188" lastFinishedPulling="2026-01-24 00:55:22.939290551 +0000 UTC m=+3106.969261562" observedRunningTime="2026-01-24 00:55:23.961835299 +0000 UTC m=+3107.991806350" watchObservedRunningTime="2026-01-24 00:55:23.963461929 +0000 UTC m=+3107.993432960" Jan 24 00:55:29 crc kubenswrapper[4676]: I0124 00:55:29.256353 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:55:29 crc kubenswrapper[4676]: E0124 00:55:29.258119 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:55:44 crc kubenswrapper[4676]: I0124 00:55:44.268718 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:55:44 crc kubenswrapper[4676]: E0124 00:55:44.269957 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.577546 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-57xmz/must-gather-hv4gr"] Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.579293 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.582328 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-57xmz"/"kube-root-ca.crt" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.582954 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-57xmz"/"openshift-service-ca.crt" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.598636 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-57xmz/must-gather-hv4gr"] Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.711546 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a646626a-4429-4d8a-9f75-7b8bfade90ec-must-gather-output\") pod \"must-gather-hv4gr\" (UID: \"a646626a-4429-4d8a-9f75-7b8bfade90ec\") " pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.711686 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbztc\" (UniqueName: \"kubernetes.io/projected/a646626a-4429-4d8a-9f75-7b8bfade90ec-kube-api-access-jbztc\") pod \"must-gather-hv4gr\" (UID: \"a646626a-4429-4d8a-9f75-7b8bfade90ec\") " pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.813963 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a646626a-4429-4d8a-9f75-7b8bfade90ec-must-gather-output\") pod \"must-gather-hv4gr\" (UID: \"a646626a-4429-4d8a-9f75-7b8bfade90ec\") " pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.814024 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbztc\" (UniqueName: \"kubernetes.io/projected/a646626a-4429-4d8a-9f75-7b8bfade90ec-kube-api-access-jbztc\") pod \"must-gather-hv4gr\" (UID: \"a646626a-4429-4d8a-9f75-7b8bfade90ec\") " pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.814422 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a646626a-4429-4d8a-9f75-7b8bfade90ec-must-gather-output\") pod \"must-gather-hv4gr\" (UID: \"a646626a-4429-4d8a-9f75-7b8bfade90ec\") " pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.845057 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbztc\" (UniqueName: \"kubernetes.io/projected/a646626a-4429-4d8a-9f75-7b8bfade90ec-kube-api-access-jbztc\") pod \"must-gather-hv4gr\" (UID: \"a646626a-4429-4d8a-9f75-7b8bfade90ec\") " pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 00:55:45 crc kubenswrapper[4676]: I0124 00:55:45.896945 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 00:55:46 crc kubenswrapper[4676]: I0124 00:55:46.335125 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-57xmz/must-gather-hv4gr"] Jan 24 00:55:47 crc kubenswrapper[4676]: I0124 00:55:47.174855 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57xmz/must-gather-hv4gr" event={"ID":"a646626a-4429-4d8a-9f75-7b8bfade90ec","Type":"ContainerStarted","Data":"43c717b44f07040232b60ea2476de8679d1451b8e216e87add1868ef3142ef40"} Jan 24 00:55:54 crc kubenswrapper[4676]: I0124 00:55:54.246611 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57xmz/must-gather-hv4gr" event={"ID":"a646626a-4429-4d8a-9f75-7b8bfade90ec","Type":"ContainerStarted","Data":"5282e9521531b7fbcb12f5cd407d932cd2321b1dd6ad9e7d8b88355c7f251d0b"} Jan 24 00:55:54 crc kubenswrapper[4676]: I0124 00:55:54.247264 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57xmz/must-gather-hv4gr" event={"ID":"a646626a-4429-4d8a-9f75-7b8bfade90ec","Type":"ContainerStarted","Data":"e762fbd56fb888c091ce2f8ff2b6ad8a52059af7b071066cce8729c1ec864419"} Jan 24 00:55:54 crc kubenswrapper[4676]: I0124 00:55:54.282986 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-57xmz/must-gather-hv4gr" podStartSLOduration=2.456964465 podStartE2EDuration="9.282965963s" podCreationTimestamp="2026-01-24 00:55:45 +0000 UTC" firstStartedPulling="2026-01-24 00:55:46.343482517 +0000 UTC m=+3130.373453518" lastFinishedPulling="2026-01-24 00:55:53.169484015 +0000 UTC m=+3137.199455016" observedRunningTime="2026-01-24 00:55:54.272708298 +0000 UTC m=+3138.302679299" watchObservedRunningTime="2026-01-24 00:55:54.282965963 +0000 UTC m=+3138.312936974" Jan 24 00:55:57 crc kubenswrapper[4676]: I0124 00:55:57.841029 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-57xmz/crc-debug-549dl"] Jan 24 00:55:57 crc kubenswrapper[4676]: I0124 00:55:57.842504 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:55:57 crc kubenswrapper[4676]: I0124 00:55:57.845196 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-57xmz"/"default-dockercfg-qwqnc" Jan 24 00:55:57 crc kubenswrapper[4676]: I0124 00:55:57.956842 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsvs9\" (UniqueName: \"kubernetes.io/projected/fffc7981-c6a6-4477-8d51-b33a0fc21549-kube-api-access-hsvs9\") pod \"crc-debug-549dl\" (UID: \"fffc7981-c6a6-4477-8d51-b33a0fc21549\") " pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:55:57 crc kubenswrapper[4676]: I0124 00:55:57.956944 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fffc7981-c6a6-4477-8d51-b33a0fc21549-host\") pod \"crc-debug-549dl\" (UID: \"fffc7981-c6a6-4477-8d51-b33a0fc21549\") " pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:55:58 crc kubenswrapper[4676]: I0124 00:55:58.061358 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsvs9\" (UniqueName: \"kubernetes.io/projected/fffc7981-c6a6-4477-8d51-b33a0fc21549-kube-api-access-hsvs9\") pod \"crc-debug-549dl\" (UID: \"fffc7981-c6a6-4477-8d51-b33a0fc21549\") " pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:55:58 crc kubenswrapper[4676]: I0124 00:55:58.061458 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fffc7981-c6a6-4477-8d51-b33a0fc21549-host\") pod \"crc-debug-549dl\" (UID: \"fffc7981-c6a6-4477-8d51-b33a0fc21549\") " pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:55:58 crc kubenswrapper[4676]: I0124 00:55:58.061570 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fffc7981-c6a6-4477-8d51-b33a0fc21549-host\") pod \"crc-debug-549dl\" (UID: \"fffc7981-c6a6-4477-8d51-b33a0fc21549\") " pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:55:58 crc kubenswrapper[4676]: I0124 00:55:58.084867 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsvs9\" (UniqueName: \"kubernetes.io/projected/fffc7981-c6a6-4477-8d51-b33a0fc21549-kube-api-access-hsvs9\") pod \"crc-debug-549dl\" (UID: \"fffc7981-c6a6-4477-8d51-b33a0fc21549\") " pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:55:58 crc kubenswrapper[4676]: I0124 00:55:58.162012 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:55:58 crc kubenswrapper[4676]: W0124 00:55:58.204125 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfffc7981_c6a6_4477_8d51_b33a0fc21549.slice/crio-7f3d83976e681330726b254205952d75eef12938a4edbadd8334f9e1893a505d WatchSource:0}: Error finding container 7f3d83976e681330726b254205952d75eef12938a4edbadd8334f9e1893a505d: Status 404 returned error can't find the container with id 7f3d83976e681330726b254205952d75eef12938a4edbadd8334f9e1893a505d Jan 24 00:55:58 crc kubenswrapper[4676]: I0124 00:55:58.255927 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:55:58 crc kubenswrapper[4676]: E0124 00:55:58.256155 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:55:58 crc kubenswrapper[4676]: I0124 00:55:58.284292 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57xmz/crc-debug-549dl" event={"ID":"fffc7981-c6a6-4477-8d51-b33a0fc21549","Type":"ContainerStarted","Data":"7f3d83976e681330726b254205952d75eef12938a4edbadd8334f9e1893a505d"} Jan 24 00:56:10 crc kubenswrapper[4676]: I0124 00:56:10.255826 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:56:10 crc kubenswrapper[4676]: E0124 00:56:10.256629 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:56:11 crc kubenswrapper[4676]: I0124 00:56:11.397700 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57xmz/crc-debug-549dl" event={"ID":"fffc7981-c6a6-4477-8d51-b33a0fc21549","Type":"ContainerStarted","Data":"fb528b96bb0ea9229c8a0baf4db69e4aad9f43be8642c6a3de48a10855ae6586"} Jan 24 00:56:11 crc kubenswrapper[4676]: I0124 00:56:11.418545 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-57xmz/crc-debug-549dl" podStartSLOduration=2.319340692 podStartE2EDuration="14.41852672s" podCreationTimestamp="2026-01-24 00:55:57 +0000 UTC" firstStartedPulling="2026-01-24 00:55:58.205579855 +0000 UTC m=+3142.235550856" lastFinishedPulling="2026-01-24 00:56:10.304765883 +0000 UTC m=+3154.334736884" observedRunningTime="2026-01-24 00:56:11.414854367 +0000 UTC m=+3155.444825378" watchObservedRunningTime="2026-01-24 00:56:11.41852672 +0000 UTC m=+3155.448497731" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.212958 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hkwvj"] Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.216104 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.226358 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkwvj"] Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.387402 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-utilities\") pod \"community-operators-hkwvj\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.387514 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvv8z\" (UniqueName: \"kubernetes.io/projected/57b3d882-428b-47c9-97e6-9c26f4334f3e-kube-api-access-lvv8z\") pod \"community-operators-hkwvj\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.388266 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-catalog-content\") pod \"community-operators-hkwvj\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.490787 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-utilities\") pod \"community-operators-hkwvj\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.490844 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvv8z\" (UniqueName: \"kubernetes.io/projected/57b3d882-428b-47c9-97e6-9c26f4334f3e-kube-api-access-lvv8z\") pod \"community-operators-hkwvj\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.491640 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-utilities\") pod \"community-operators-hkwvj\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.491980 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-catalog-content\") pod \"community-operators-hkwvj\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.492418 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-catalog-content\") pod \"community-operators-hkwvj\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.522971 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvv8z\" (UniqueName: \"kubernetes.io/projected/57b3d882-428b-47c9-97e6-9c26f4334f3e-kube-api-access-lvv8z\") pod \"community-operators-hkwvj\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:17 crc kubenswrapper[4676]: I0124 00:56:17.564619 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:19 crc kubenswrapper[4676]: I0124 00:56:19.021464 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hkwvj"] Jan 24 00:56:19 crc kubenswrapper[4676]: I0124 00:56:19.466133 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkwvj" event={"ID":"57b3d882-428b-47c9-97e6-9c26f4334f3e","Type":"ContainerStarted","Data":"a0eed663c019c0254d4baf55709bf2bd4422bac18c67da9b1501732b33846d24"} Jan 24 00:56:20 crc kubenswrapper[4676]: I0124 00:56:20.475212 4676 generic.go:334] "Generic (PLEG): container finished" podID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerID="425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a" exitCode=0 Jan 24 00:56:20 crc kubenswrapper[4676]: I0124 00:56:20.475389 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkwvj" event={"ID":"57b3d882-428b-47c9-97e6-9c26f4334f3e","Type":"ContainerDied","Data":"425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a"} Jan 24 00:56:21 crc kubenswrapper[4676]: I0124 00:56:21.486216 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkwvj" event={"ID":"57b3d882-428b-47c9-97e6-9c26f4334f3e","Type":"ContainerStarted","Data":"5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58"} Jan 24 00:56:22 crc kubenswrapper[4676]: I0124 00:56:22.500156 4676 generic.go:334] "Generic (PLEG): container finished" podID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerID="5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58" exitCode=0 Jan 24 00:56:22 crc kubenswrapper[4676]: I0124 00:56:22.500235 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkwvj" event={"ID":"57b3d882-428b-47c9-97e6-9c26f4334f3e","Type":"ContainerDied","Data":"5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58"} Jan 24 00:56:25 crc kubenswrapper[4676]: I0124 00:56:25.257030 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:56:25 crc kubenswrapper[4676]: E0124 00:56:25.257672 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:56:25 crc kubenswrapper[4676]: I0124 00:56:25.524681 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkwvj" event={"ID":"57b3d882-428b-47c9-97e6-9c26f4334f3e","Type":"ContainerStarted","Data":"00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127"} Jan 24 00:56:25 crc kubenswrapper[4676]: I0124 00:56:25.544631 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hkwvj" podStartSLOduration=5.979579741 podStartE2EDuration="8.544618059s" podCreationTimestamp="2026-01-24 00:56:17 +0000 UTC" firstStartedPulling="2026-01-24 00:56:20.477515417 +0000 UTC m=+3164.507486418" lastFinishedPulling="2026-01-24 00:56:23.042553735 +0000 UTC m=+3167.072524736" observedRunningTime="2026-01-24 00:56:25.541188563 +0000 UTC m=+3169.571159564" watchObservedRunningTime="2026-01-24 00:56:25.544618059 +0000 UTC m=+3169.574589060" Jan 24 00:56:27 crc kubenswrapper[4676]: I0124 00:56:27.547467 4676 generic.go:334] "Generic (PLEG): container finished" podID="fffc7981-c6a6-4477-8d51-b33a0fc21549" containerID="fb528b96bb0ea9229c8a0baf4db69e4aad9f43be8642c6a3de48a10855ae6586" exitCode=0 Jan 24 00:56:27 crc kubenswrapper[4676]: I0124 00:56:27.547654 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57xmz/crc-debug-549dl" event={"ID":"fffc7981-c6a6-4477-8d51-b33a0fc21549","Type":"ContainerDied","Data":"fb528b96bb0ea9229c8a0baf4db69e4aad9f43be8642c6a3de48a10855ae6586"} Jan 24 00:56:27 crc kubenswrapper[4676]: I0124 00:56:27.565396 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:27 crc kubenswrapper[4676]: I0124 00:56:27.566480 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:27 crc kubenswrapper[4676]: I0124 00:56:27.614581 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:28 crc kubenswrapper[4676]: I0124 00:56:28.643097 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:56:28 crc kubenswrapper[4676]: I0124 00:56:28.681676 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-57xmz/crc-debug-549dl"] Jan 24 00:56:28 crc kubenswrapper[4676]: I0124 00:56:28.694723 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-57xmz/crc-debug-549dl"] Jan 24 00:56:28 crc kubenswrapper[4676]: I0124 00:56:28.841495 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fffc7981-c6a6-4477-8d51-b33a0fc21549-host\") pod \"fffc7981-c6a6-4477-8d51-b33a0fc21549\" (UID: \"fffc7981-c6a6-4477-8d51-b33a0fc21549\") " Jan 24 00:56:28 crc kubenswrapper[4676]: I0124 00:56:28.841668 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsvs9\" (UniqueName: \"kubernetes.io/projected/fffc7981-c6a6-4477-8d51-b33a0fc21549-kube-api-access-hsvs9\") pod \"fffc7981-c6a6-4477-8d51-b33a0fc21549\" (UID: \"fffc7981-c6a6-4477-8d51-b33a0fc21549\") " Jan 24 00:56:28 crc kubenswrapper[4676]: I0124 00:56:28.841749 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fffc7981-c6a6-4477-8d51-b33a0fc21549-host" (OuterVolumeSpecName: "host") pod "fffc7981-c6a6-4477-8d51-b33a0fc21549" (UID: "fffc7981-c6a6-4477-8d51-b33a0fc21549"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:56:28 crc kubenswrapper[4676]: I0124 00:56:28.842159 4676 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fffc7981-c6a6-4477-8d51-b33a0fc21549-host\") on node \"crc\" DevicePath \"\"" Jan 24 00:56:28 crc kubenswrapper[4676]: I0124 00:56:28.848617 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fffc7981-c6a6-4477-8d51-b33a0fc21549-kube-api-access-hsvs9" (OuterVolumeSpecName: "kube-api-access-hsvs9") pod "fffc7981-c6a6-4477-8d51-b33a0fc21549" (UID: "fffc7981-c6a6-4477-8d51-b33a0fc21549"). InnerVolumeSpecName "kube-api-access-hsvs9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:56:28 crc kubenswrapper[4676]: I0124 00:56:28.943823 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsvs9\" (UniqueName: \"kubernetes.io/projected/fffc7981-c6a6-4477-8d51-b33a0fc21549-kube-api-access-hsvs9\") on node \"crc\" DevicePath \"\"" Jan 24 00:56:29 crc kubenswrapper[4676]: I0124 00:56:29.565316 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/crc-debug-549dl" Jan 24 00:56:29 crc kubenswrapper[4676]: I0124 00:56:29.566581 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f3d83976e681330726b254205952d75eef12938a4edbadd8334f9e1893a505d" Jan 24 00:56:29 crc kubenswrapper[4676]: I0124 00:56:29.640168 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:29 crc kubenswrapper[4676]: I0124 00:56:29.691960 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkwvj"] Jan 24 00:56:29 crc kubenswrapper[4676]: I0124 00:56:29.878231 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-57xmz/crc-debug-xdbwf"] Jan 24 00:56:29 crc kubenswrapper[4676]: E0124 00:56:29.878641 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fffc7981-c6a6-4477-8d51-b33a0fc21549" containerName="container-00" Jan 24 00:56:29 crc kubenswrapper[4676]: I0124 00:56:29.878658 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="fffc7981-c6a6-4477-8d51-b33a0fc21549" containerName="container-00" Jan 24 00:56:29 crc kubenswrapper[4676]: I0124 00:56:29.878808 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="fffc7981-c6a6-4477-8d51-b33a0fc21549" containerName="container-00" Jan 24 00:56:29 crc kubenswrapper[4676]: I0124 00:56:29.879366 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:29 crc kubenswrapper[4676]: I0124 00:56:29.882435 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-57xmz"/"default-dockercfg-qwqnc" Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.065670 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e133ba60-b33b-4289-a6a7-ef3827079dbb-host\") pod \"crc-debug-xdbwf\" (UID: \"e133ba60-b33b-4289-a6a7-ef3827079dbb\") " pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.065788 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcfd9\" (UniqueName: \"kubernetes.io/projected/e133ba60-b33b-4289-a6a7-ef3827079dbb-kube-api-access-xcfd9\") pod \"crc-debug-xdbwf\" (UID: \"e133ba60-b33b-4289-a6a7-ef3827079dbb\") " pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.167295 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcfd9\" (UniqueName: \"kubernetes.io/projected/e133ba60-b33b-4289-a6a7-ef3827079dbb-kube-api-access-xcfd9\") pod \"crc-debug-xdbwf\" (UID: \"e133ba60-b33b-4289-a6a7-ef3827079dbb\") " pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.167432 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e133ba60-b33b-4289-a6a7-ef3827079dbb-host\") pod \"crc-debug-xdbwf\" (UID: \"e133ba60-b33b-4289-a6a7-ef3827079dbb\") " pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.167540 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e133ba60-b33b-4289-a6a7-ef3827079dbb-host\") pod \"crc-debug-xdbwf\" (UID: \"e133ba60-b33b-4289-a6a7-ef3827079dbb\") " pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.190546 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcfd9\" (UniqueName: \"kubernetes.io/projected/e133ba60-b33b-4289-a6a7-ef3827079dbb-kube-api-access-xcfd9\") pod \"crc-debug-xdbwf\" (UID: \"e133ba60-b33b-4289-a6a7-ef3827079dbb\") " pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.202090 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:30 crc kubenswrapper[4676]: W0124 00:56:30.255839 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode133ba60_b33b_4289_a6a7_ef3827079dbb.slice/crio-a2bff01c8277b0be8b9a906d9b3e4b0421a15dcefbe34a184e886e13d3684e31 WatchSource:0}: Error finding container a2bff01c8277b0be8b9a906d9b3e4b0421a15dcefbe34a184e886e13d3684e31: Status 404 returned error can't find the container with id a2bff01c8277b0be8b9a906d9b3e4b0421a15dcefbe34a184e886e13d3684e31 Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.273210 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fffc7981-c6a6-4477-8d51-b33a0fc21549" path="/var/lib/kubelet/pods/fffc7981-c6a6-4477-8d51-b33a0fc21549/volumes" Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.572155 4676 generic.go:334] "Generic (PLEG): container finished" podID="e133ba60-b33b-4289-a6a7-ef3827079dbb" containerID="c2d2cba815cc8f80eb4e37a292a7de3ac1d314214ffa71dbce141268908119a0" exitCode=1 Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.572258 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57xmz/crc-debug-xdbwf" event={"ID":"e133ba60-b33b-4289-a6a7-ef3827079dbb","Type":"ContainerDied","Data":"c2d2cba815cc8f80eb4e37a292a7de3ac1d314214ffa71dbce141268908119a0"} Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.572471 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57xmz/crc-debug-xdbwf" event={"ID":"e133ba60-b33b-4289-a6a7-ef3827079dbb","Type":"ContainerStarted","Data":"a2bff01c8277b0be8b9a906d9b3e4b0421a15dcefbe34a184e886e13d3684e31"} Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.609477 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-57xmz/crc-debug-xdbwf"] Jan 24 00:56:30 crc kubenswrapper[4676]: I0124 00:56:30.619603 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-57xmz/crc-debug-xdbwf"] Jan 24 00:56:31 crc kubenswrapper[4676]: I0124 00:56:31.579394 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hkwvj" podUID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerName="registry-server" containerID="cri-o://00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127" gracePeriod=2 Jan 24 00:56:31 crc kubenswrapper[4676]: I0124 00:56:31.752157 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:31 crc kubenswrapper[4676]: I0124 00:56:31.910439 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e133ba60-b33b-4289-a6a7-ef3827079dbb-host\") pod \"e133ba60-b33b-4289-a6a7-ef3827079dbb\" (UID: \"e133ba60-b33b-4289-a6a7-ef3827079dbb\") " Jan 24 00:56:31 crc kubenswrapper[4676]: I0124 00:56:31.910558 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcfd9\" (UniqueName: \"kubernetes.io/projected/e133ba60-b33b-4289-a6a7-ef3827079dbb-kube-api-access-xcfd9\") pod \"e133ba60-b33b-4289-a6a7-ef3827079dbb\" (UID: \"e133ba60-b33b-4289-a6a7-ef3827079dbb\") " Jan 24 00:56:31 crc kubenswrapper[4676]: I0124 00:56:31.910578 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e133ba60-b33b-4289-a6a7-ef3827079dbb-host" (OuterVolumeSpecName: "host") pod "e133ba60-b33b-4289-a6a7-ef3827079dbb" (UID: "e133ba60-b33b-4289-a6a7-ef3827079dbb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 00:56:31 crc kubenswrapper[4676]: I0124 00:56:31.911110 4676 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e133ba60-b33b-4289-a6a7-ef3827079dbb-host\") on node \"crc\" DevicePath \"\"" Jan 24 00:56:31 crc kubenswrapper[4676]: I0124 00:56:31.923523 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e133ba60-b33b-4289-a6a7-ef3827079dbb-kube-api-access-xcfd9" (OuterVolumeSpecName: "kube-api-access-xcfd9") pod "e133ba60-b33b-4289-a6a7-ef3827079dbb" (UID: "e133ba60-b33b-4289-a6a7-ef3827079dbb"). InnerVolumeSpecName "kube-api-access-xcfd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.007264 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.012796 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvv8z\" (UniqueName: \"kubernetes.io/projected/57b3d882-428b-47c9-97e6-9c26f4334f3e-kube-api-access-lvv8z\") pod \"57b3d882-428b-47c9-97e6-9c26f4334f3e\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.012970 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-utilities\") pod \"57b3d882-428b-47c9-97e6-9c26f4334f3e\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.012995 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-catalog-content\") pod \"57b3d882-428b-47c9-97e6-9c26f4334f3e\" (UID: \"57b3d882-428b-47c9-97e6-9c26f4334f3e\") " Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.013471 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcfd9\" (UniqueName: \"kubernetes.io/projected/e133ba60-b33b-4289-a6a7-ef3827079dbb-kube-api-access-xcfd9\") on node \"crc\" DevicePath \"\"" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.013950 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-utilities" (OuterVolumeSpecName: "utilities") pod "57b3d882-428b-47c9-97e6-9c26f4334f3e" (UID: "57b3d882-428b-47c9-97e6-9c26f4334f3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.016116 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b3d882-428b-47c9-97e6-9c26f4334f3e-kube-api-access-lvv8z" (OuterVolumeSpecName: "kube-api-access-lvv8z") pod "57b3d882-428b-47c9-97e6-9c26f4334f3e" (UID: "57b3d882-428b-47c9-97e6-9c26f4334f3e"). InnerVolumeSpecName "kube-api-access-lvv8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.091830 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57b3d882-428b-47c9-97e6-9c26f4334f3e" (UID: "57b3d882-428b-47c9-97e6-9c26f4334f3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.115789 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.115821 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57b3d882-428b-47c9-97e6-9c26f4334f3e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.115833 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvv8z\" (UniqueName: \"kubernetes.io/projected/57b3d882-428b-47c9-97e6-9c26f4334f3e-kube-api-access-lvv8z\") on node \"crc\" DevicePath \"\"" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.268965 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e133ba60-b33b-4289-a6a7-ef3827079dbb" path="/var/lib/kubelet/pods/e133ba60-b33b-4289-a6a7-ef3827079dbb/volumes" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.598394 4676 scope.go:117] "RemoveContainer" containerID="c2d2cba815cc8f80eb4e37a292a7de3ac1d314214ffa71dbce141268908119a0" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.598413 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/crc-debug-xdbwf" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.606715 4676 generic.go:334] "Generic (PLEG): container finished" podID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerID="00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127" exitCode=0 Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.606825 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkwvj" event={"ID":"57b3d882-428b-47c9-97e6-9c26f4334f3e","Type":"ContainerDied","Data":"00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127"} Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.606887 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hkwvj" event={"ID":"57b3d882-428b-47c9-97e6-9c26f4334f3e","Type":"ContainerDied","Data":"a0eed663c019c0254d4baf55709bf2bd4422bac18c67da9b1501732b33846d24"} Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.607102 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hkwvj" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.656672 4676 scope.go:117] "RemoveContainer" containerID="00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.704293 4676 scope.go:117] "RemoveContainer" containerID="5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.708061 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hkwvj"] Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.718468 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hkwvj"] Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.747857 4676 scope.go:117] "RemoveContainer" containerID="425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.770990 4676 scope.go:117] "RemoveContainer" containerID="00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127" Jan 24 00:56:32 crc kubenswrapper[4676]: E0124 00:56:32.774498 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127\": container with ID starting with 00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127 not found: ID does not exist" containerID="00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.774552 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127"} err="failed to get container status \"00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127\": rpc error: code = NotFound desc = could not find container \"00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127\": container with ID starting with 00383a41a004aac9747fb6b4d29966344c64ce12b2249d555c82a587d77fb127 not found: ID does not exist" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.774586 4676 scope.go:117] "RemoveContainer" containerID="5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58" Jan 24 00:56:32 crc kubenswrapper[4676]: E0124 00:56:32.774918 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58\": container with ID starting with 5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58 not found: ID does not exist" containerID="5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.774952 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58"} err="failed to get container status \"5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58\": rpc error: code = NotFound desc = could not find container \"5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58\": container with ID starting with 5381f8e56de8f59e8cd953bea2d31650286f7d0073378850f8e97c4bf617cb58 not found: ID does not exist" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.774976 4676 scope.go:117] "RemoveContainer" containerID="425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a" Jan 24 00:56:32 crc kubenswrapper[4676]: E0124 00:56:32.775254 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a\": container with ID starting with 425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a not found: ID does not exist" containerID="425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a" Jan 24 00:56:32 crc kubenswrapper[4676]: I0124 00:56:32.775280 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a"} err="failed to get container status \"425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a\": rpc error: code = NotFound desc = could not find container \"425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a\": container with ID starting with 425d5cde97e1698c9bba080ab311dd33c110e8b51f818c03479ea99d3bf0e87a not found: ID does not exist" Jan 24 00:56:34 crc kubenswrapper[4676]: I0124 00:56:34.267689 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b3d882-428b-47c9-97e6-9c26f4334f3e" path="/var/lib/kubelet/pods/57b3d882-428b-47c9-97e6-9c26f4334f3e/volumes" Jan 24 00:56:40 crc kubenswrapper[4676]: I0124 00:56:40.256175 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:56:40 crc kubenswrapper[4676]: E0124 00:56:40.257090 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:56:55 crc kubenswrapper[4676]: I0124 00:56:55.255668 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:56:55 crc kubenswrapper[4676]: E0124 00:56:55.256451 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:57:10 crc kubenswrapper[4676]: I0124 00:57:10.255860 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:57:10 crc kubenswrapper[4676]: E0124 00:57:10.256571 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:57:18 crc kubenswrapper[4676]: I0124 00:57:18.971911 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54b7b855f4-s49zw_c015646b-9361-4b5c-b465-a66d7fc5cc53/barbican-api/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.121185 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54b7b855f4-s49zw_c015646b-9361-4b5c-b465-a66d7fc5cc53/barbican-api-log/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.185823 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79ccd69c74-nj8k6_a05052ce-062e-423c-80cf-78349e42718f/barbican-keystone-listener/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.247650 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79ccd69c74-nj8k6_a05052ce-062e-423c-80cf-78349e42718f/barbican-keystone-listener-log/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.401604 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68bd7fb46c-sflbz_009f35e0-3a98-453b-b92e-8db9e5c92798/barbican-worker-log/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.410070 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68bd7fb46c-sflbz_009f35e0-3a98-453b-b92e-8db9e5c92798/barbican-worker/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.613815 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp_1fd4e1f4-0772-493a-b929-6e93470f9abf/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.725126 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58bbd77a-9518-4037-b96b-a1490082fb04/ceilometer-notification-agent/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.734104 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58bbd77a-9518-4037-b96b-a1490082fb04/ceilometer-central-agent/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.841206 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58bbd77a-9518-4037-b96b-a1490082fb04/proxy-httpd/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.940752 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d38cef52-387b-4633-be75-6dc455ad53c4/cinder-api/0.log" Jan 24 00:57:19 crc kubenswrapper[4676]: I0124 00:57:19.972597 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58bbd77a-9518-4037-b96b-a1490082fb04/sg-core/0.log" Jan 24 00:57:20 crc kubenswrapper[4676]: I0124 00:57:20.079937 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d38cef52-387b-4633-be75-6dc455ad53c4/cinder-api-log/0.log" Jan 24 00:57:20 crc kubenswrapper[4676]: I0124 00:57:20.231039 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978/cinder-scheduler/0.log" Jan 24 00:57:20 crc kubenswrapper[4676]: I0124 00:57:20.279063 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978/probe/0.log" Jan 24 00:57:20 crc kubenswrapper[4676]: I0124 00:57:20.430103 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg_7f552002-93ef-485f-9227-a94733534466/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:20 crc kubenswrapper[4676]: I0124 00:57:20.568107 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7_f6bc5ee4-f730-4e1e-9684-b643daed2519/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:20 crc kubenswrapper[4676]: I0124 00:57:20.677499 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-drclt_cdf394f7-5d67-4a0f-9644-82fe83a72e2d/init/0.log" Jan 24 00:57:20 crc kubenswrapper[4676]: I0124 00:57:20.840723 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-drclt_cdf394f7-5d67-4a0f-9644-82fe83a72e2d/init/0.log" Jan 24 00:57:20 crc kubenswrapper[4676]: I0124 00:57:20.934723 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-drclt_cdf394f7-5d67-4a0f-9644-82fe83a72e2d/dnsmasq-dns/0.log" Jan 24 00:57:20 crc kubenswrapper[4676]: I0124 00:57:20.964822 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-skh8k_7e20d58a-5f01-4c23-9ab0-650d3ae76844/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:21 crc kubenswrapper[4676]: I0124 00:57:21.127085 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a6cef518-8385-4315-83d5-f46a6144d5a0/glance-httpd/0.log" Jan 24 00:57:21 crc kubenswrapper[4676]: I0124 00:57:21.138907 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a6cef518-8385-4315-83d5-f46a6144d5a0/glance-log/0.log" Jan 24 00:57:21 crc kubenswrapper[4676]: I0124 00:57:21.298992 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_042ec835-b8c1-43be-a19d-d70f76128e26/glance-log/0.log" Jan 24 00:57:21 crc kubenswrapper[4676]: I0124 00:57:21.342328 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_042ec835-b8c1-43be-a19d-d70f76128e26/glance-httpd/0.log" Jan 24 00:57:21 crc kubenswrapper[4676]: I0124 00:57:21.483228 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f876ddf46-fs7qv_ac7dce6b-3bd9-4ad9-9485-83d9384b8bad/horizon/1.log" Jan 24 00:57:21 crc kubenswrapper[4676]: I0124 00:57:21.559811 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f876ddf46-fs7qv_ac7dce6b-3bd9-4ad9-9485-83d9384b8bad/horizon/0.log" Jan 24 00:57:21 crc kubenswrapper[4676]: I0124 00:57:21.739976 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs_e9d2a92f-e22e-44c5-86c5-8b38824e3d4c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:21 crc kubenswrapper[4676]: I0124 00:57:21.869954 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f876ddf46-fs7qv_ac7dce6b-3bd9-4ad9-9485-83d9384b8bad/horizon-log/0.log" Jan 24 00:57:21 crc kubenswrapper[4676]: I0124 00:57:21.971345 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-sz24c_15fbef8d-e606-4b72-a994-e71d03e8fec8/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:22 crc kubenswrapper[4676]: I0124 00:57:22.127131 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-849b597d57-kw79c_29bf9e4b-4fb3-41f4-9280-f7ea2e61a844/keystone-api/0.log" Jan 24 00:57:22 crc kubenswrapper[4676]: I0124 00:57:22.202326 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_25ee749c-0b84-4abd-9fe0-a6f23c0c912d/kube-state-metrics/0.log" Jan 24 00:57:22 crc kubenswrapper[4676]: I0124 00:57:22.345850 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ddk99_2cafe497-da96-4b39-bec2-1ec54f859303/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:22 crc kubenswrapper[4676]: I0124 00:57:22.978580 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd579cfd9-q4npp_6dda8455-6777-4efc-abf3-df547cf58339/neutron-api/0.log" Jan 24 00:57:23 crc kubenswrapper[4676]: I0124 00:57:23.077188 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd579cfd9-q4npp_6dda8455-6777-4efc-abf3-df547cf58339/neutron-httpd/0.log" Jan 24 00:57:23 crc kubenswrapper[4676]: I0124 00:57:23.270945 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7_cba640eb-c65f-46be-af5d-5126418c361a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:23 crc kubenswrapper[4676]: I0124 00:57:23.601334 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ff542eb3-0142-4d46-a6e6-2e89c73f5824/nova-api-log/0.log" Jan 24 00:57:23 crc kubenswrapper[4676]: I0124 00:57:23.716801 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ff542eb3-0142-4d46-a6e6-2e89c73f5824/nova-api-api/0.log" Jan 24 00:57:23 crc kubenswrapper[4676]: I0124 00:57:23.726003 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_400ba963-913a-401c-8f2e-21005977e0c2/nova-cell0-conductor-conductor/0.log" Jan 24 00:57:23 crc kubenswrapper[4676]: I0124 00:57:23.945517 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_4420d197-3908-406b-9661-67d64ccd7768/nova-cell1-conductor-conductor/0.log" Jan 24 00:57:24 crc kubenswrapper[4676]: I0124 00:57:24.095892 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_b1499ed8-1041-4980-b9da-cb957cbf215c/nova-cell1-novncproxy-novncproxy/0.log" Jan 24 00:57:24 crc kubenswrapper[4676]: I0124 00:57:24.295256 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5cl2g_75cae94d-b818-4f92-b42d-fd8cec63a657/nova-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:24 crc kubenswrapper[4676]: I0124 00:57:24.464625 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1085d48-78f1-4437-a518-239ba90b7c0b/nova-metadata-log/0.log" Jan 24 00:57:24 crc kubenswrapper[4676]: I0124 00:57:24.755413 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8ff6f87b-a571-43da-9fbc-9203f0001771/nova-scheduler-scheduler/0.log" Jan 24 00:57:24 crc kubenswrapper[4676]: I0124 00:57:24.815964 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_19365292-50d8-4e94-952f-2df7ee20f0ba/mysql-bootstrap/0.log" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.068366 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_19365292-50d8-4e94-952f-2df7ee20f0ba/mysql-bootstrap/0.log" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.099531 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_19365292-50d8-4e94-952f-2df7ee20f0ba/galera/0.log" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.255201 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:57:25 crc kubenswrapper[4676]: E0124 00:57:25.255449 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.280137 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2bbaae64-ac2d-43c6-8984-5483f2eb4211/mysql-bootstrap/0.log" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.353904 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1085d48-78f1-4437-a518-239ba90b7c0b/nova-metadata-metadata/0.log" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.578317 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2bbaae64-ac2d-43c6-8984-5483f2eb4211/mysql-bootstrap/0.log" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.643031 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f3fa6e59-785d-4d40-8d73-170552068e43/openstackclient/0.log" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.650697 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2bbaae64-ac2d-43c6-8984-5483f2eb4211/galera/0.log" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.821393 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jsd4n_7adbcf83-efbd-4e8d-97e5-f8768463284a/ovn-controller/0.log" Jan 24 00:57:25 crc kubenswrapper[4676]: I0124 00:57:25.956307 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-mq6ns_21550436-cd71-46d8-838e-c51c19ddf8ff/openstack-network-exporter/0.log" Jan 24 00:57:26 crc kubenswrapper[4676]: I0124 00:57:26.151673 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6sl9q_427fbd2d-16ef-44a6-a71d-8172f56b863d/ovsdb-server-init/0.log" Jan 24 00:57:26 crc kubenswrapper[4676]: I0124 00:57:26.450044 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6sl9q_427fbd2d-16ef-44a6-a71d-8172f56b863d/ovsdb-server-init/0.log" Jan 24 00:57:26 crc kubenswrapper[4676]: I0124 00:57:26.457369 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6sl9q_427fbd2d-16ef-44a6-a71d-8172f56b863d/ovs-vswitchd/0.log" Jan 24 00:57:26 crc kubenswrapper[4676]: I0124 00:57:26.462997 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6sl9q_427fbd2d-16ef-44a6-a71d-8172f56b863d/ovsdb-server/0.log" Jan 24 00:57:26 crc kubenswrapper[4676]: I0124 00:57:26.723910 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1486ea92-d267-49cd-8516-d474ef25c2df/openstack-network-exporter/0.log" Jan 24 00:57:26 crc kubenswrapper[4676]: I0124 00:57:26.765053 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-527bv_55444bfa-a024-4606-aa57-6456c6688e52/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:26 crc kubenswrapper[4676]: I0124 00:57:26.874096 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1486ea92-d267-49cd-8516-d474ef25c2df/ovn-northd/0.log" Jan 24 00:57:27 crc kubenswrapper[4676]: I0124 00:57:27.001846 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3bff778a-b10f-4ba9-a12f-f4086608fd30/openstack-network-exporter/0.log" Jan 24 00:57:27 crc kubenswrapper[4676]: I0124 00:57:27.113216 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3bff778a-b10f-4ba9-a12f-f4086608fd30/ovsdbserver-nb/0.log" Jan 24 00:57:27 crc kubenswrapper[4676]: I0124 00:57:27.214628 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f3648c65-9fcb-4a9e-b4cb-d8437dc00141/openstack-network-exporter/0.log" Jan 24 00:57:27 crc kubenswrapper[4676]: I0124 00:57:27.290580 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f3648c65-9fcb-4a9e-b4cb-d8437dc00141/ovsdbserver-sb/0.log" Jan 24 00:57:27 crc kubenswrapper[4676]: I0124 00:57:27.438486 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-765f6cdf58-5q9v9_21e63383-223a-4247-8589-03ab5a33f980/placement-api/0.log" Jan 24 00:57:27 crc kubenswrapper[4676]: I0124 00:57:27.566421 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-765f6cdf58-5q9v9_21e63383-223a-4247-8589-03ab5a33f980/placement-log/0.log" Jan 24 00:57:27 crc kubenswrapper[4676]: I0124 00:57:27.607662 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a2df1d42-fa93-4771-ba77-1c27f820b298/setup-container/0.log" Jan 24 00:57:27 crc kubenswrapper[4676]: I0124 00:57:27.865941 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a2df1d42-fa93-4771-ba77-1c27f820b298/setup-container/0.log" Jan 24 00:57:27 crc kubenswrapper[4676]: I0124 00:57:27.933014 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a2df1d42-fa93-4771-ba77-1c27f820b298/rabbitmq/0.log" Jan 24 00:57:28 crc kubenswrapper[4676]: I0124 00:57:28.020624 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c162e478-58e3-4a83-97cb-29887613c1aa/setup-container/0.log" Jan 24 00:57:28 crc kubenswrapper[4676]: I0124 00:57:28.198359 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c162e478-58e3-4a83-97cb-29887613c1aa/setup-container/0.log" Jan 24 00:57:28 crc kubenswrapper[4676]: I0124 00:57:28.307116 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c162e478-58e3-4a83-97cb-29887613c1aa/rabbitmq/0.log" Jan 24 00:57:28 crc kubenswrapper[4676]: I0124 00:57:28.318266 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-448ht_447c1e1f-d798-4bcc-a8ef-91d4ad5426a5/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:28 crc kubenswrapper[4676]: I0124 00:57:28.607908 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qx6r8_19b712ec-28a7-419f-9f09-7d0b0ecbf747/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:28 crc kubenswrapper[4676]: I0124 00:57:28.624390 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd_15cfbdc6-1f3b-49a5-8822-c4af1e686731/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:28 crc kubenswrapper[4676]: I0124 00:57:28.947951 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-mhrvn_2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:28 crc kubenswrapper[4676]: I0124 00:57:28.970170 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8b9rz_d1ed73dc-4392-4d20-a592-4a8c5ba9c104/ssh-known-hosts-edpm-deployment/0.log" Jan 24 00:57:29 crc kubenswrapper[4676]: I0124 00:57:29.265978 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f47776b4c-v4xb2_b976b9e2-b80e-4626-919d-3bb84f0151e8/proxy-server/0.log" Jan 24 00:57:29 crc kubenswrapper[4676]: I0124 00:57:29.379245 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f47776b4c-v4xb2_b976b9e2-b80e-4626-919d-3bb84f0151e8/proxy-httpd/0.log" Jan 24 00:57:29 crc kubenswrapper[4676]: I0124 00:57:29.432293 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-5fmzb_4ff61f48-e451-47e8-adcc-0870b29d28a9/swift-ring-rebalance/0.log" Jan 24 00:57:29 crc kubenswrapper[4676]: I0124 00:57:29.633434 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/account-reaper/0.log" Jan 24 00:57:29 crc kubenswrapper[4676]: I0124 00:57:29.674408 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/account-auditor/0.log" Jan 24 00:57:29 crc kubenswrapper[4676]: I0124 00:57:29.696999 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/account-replicator/0.log" Jan 24 00:57:29 crc kubenswrapper[4676]: I0124 00:57:29.830554 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/container-auditor/0.log" Jan 24 00:57:29 crc kubenswrapper[4676]: I0124 00:57:29.900091 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/account-server/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.060159 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/container-replicator/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.064485 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/container-server/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.336322 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/container-updater/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.353024 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-auditor/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.470174 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-expirer/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.489558 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-replicator/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.673676 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-updater/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.688432 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/rsync/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.714073 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-server/0.log" Jan 24 00:57:30 crc kubenswrapper[4676]: I0124 00:57:30.832330 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/swift-recon-cron/0.log" Jan 24 00:57:31 crc kubenswrapper[4676]: I0124 00:57:31.086902 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv_7ef31551-e4ed-48d0-a4d6-f9c2fb515966/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:31 crc kubenswrapper[4676]: I0124 00:57:31.119548 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_3e2adf44-9053-4dcd-9d47-27910710dbc8/tempest-tests-tempest-tests-runner/0.log" Jan 24 00:57:31 crc kubenswrapper[4676]: I0124 00:57:31.538606 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_e1574f42-e89a-40d4-b6da-2d4ef0824916/test-operator-logs-container/0.log" Jan 24 00:57:31 crc kubenswrapper[4676]: I0124 00:57:31.640077 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd_83640815-cc06-4abe-a06f-20a1f8798609/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 00:57:38 crc kubenswrapper[4676]: I0124 00:57:38.255992 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:57:38 crc kubenswrapper[4676]: E0124 00:57:38.256644 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:57:43 crc kubenswrapper[4676]: I0124 00:57:43.716461 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_545f6045-cf2f-4d4b-91d8-227148ddd71a/memcached/0.log" Jan 24 00:57:53 crc kubenswrapper[4676]: I0124 00:57:53.255929 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:57:53 crc kubenswrapper[4676]: E0124 00:57:53.256678 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:58:01 crc kubenswrapper[4676]: I0124 00:58:01.178515 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/util/0.log" Jan 24 00:58:01 crc kubenswrapper[4676]: I0124 00:58:01.288744 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/util/0.log" Jan 24 00:58:01 crc kubenswrapper[4676]: I0124 00:58:01.342669 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/pull/0.log" Jan 24 00:58:01 crc kubenswrapper[4676]: I0124 00:58:01.342782 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/pull/0.log" Jan 24 00:58:01 crc kubenswrapper[4676]: I0124 00:58:01.564433 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/pull/0.log" Jan 24 00:58:01 crc kubenswrapper[4676]: I0124 00:58:01.571796 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/util/0.log" Jan 24 00:58:01 crc kubenswrapper[4676]: I0124 00:58:01.616998 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/extract/0.log" Jan 24 00:58:01 crc kubenswrapper[4676]: I0124 00:58:01.873046 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-69cf5d4557-c8c6m_0cd05b9f-6699-46e3-ae36-9f21352e6c8e/manager/0.log" Jan 24 00:58:01 crc kubenswrapper[4676]: I0124 00:58:01.902633 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7f86f8796f-fgplq_02123851-7d2f-477b-9c60-5a9922a0bc97/manager/0.log" Jan 24 00:58:02 crc kubenswrapper[4676]: I0124 00:58:02.003592 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-js6db_a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85/manager/0.log" Jan 24 00:58:02 crc kubenswrapper[4676]: I0124 00:58:02.114062 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-btxnv_29e4b64d-19bd-419b-9e29-7a41e6f12ae0/manager/0.log" Jan 24 00:58:02 crc kubenswrapper[4676]: I0124 00:58:02.200163 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-p2nr8_5e9cf1cb-c413-45ad-8a51-bf35407fcdfe/manager/0.log" Jan 24 00:58:02 crc kubenswrapper[4676]: I0124 00:58:02.347385 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-rbzsj_6b8541f9-a37a-41d6-8006-3d0335c3abb5/manager/0.log" Jan 24 00:58:02 crc kubenswrapper[4676]: I0124 00:58:02.660903 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-598f7747c9-v25h4_555ebb8f-1bc3-4b8d-9f37-cad92b48477c/manager/0.log" Jan 24 00:58:02 crc kubenswrapper[4676]: I0124 00:58:02.667075 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-58749ffdfb-jxx26_921e121c-5261-4fe7-8171-6b634babedf4/manager/0.log" Jan 24 00:58:02 crc kubenswrapper[4676]: I0124 00:58:02.809095 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-2gzqg_9a3f9a14-1138-425d-8a56-454b282d7d9f/manager/0.log" Jan 24 00:58:02 crc kubenswrapper[4676]: I0124 00:58:02.919069 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-g4d8z_dfc79179-d245-4360-be6e-8b43441e23ed/manager/0.log" Jan 24 00:58:03 crc kubenswrapper[4676]: I0124 00:58:03.102152 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n_4ce661f6-26e2-4da2-a759-e493a60587b2/manager/0.log" Jan 24 00:58:03 crc kubenswrapper[4676]: I0124 00:58:03.140994 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78d58447c5-wqbcz_dd6346d8-9cf1-4364-b480-f4c2d872472f/manager/0.log" Jan 24 00:58:03 crc kubenswrapper[4676]: I0124 00:58:03.345277 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b8bc8d87d-4xp45_df9ab5f0-f577-4303-8045-f960c67a6936/manager/0.log" Jan 24 00:58:03 crc kubenswrapper[4676]: I0124 00:58:03.362992 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7bd9774b6-k8lw7_ced74bcb-8345-40c5-b2d4-3d369f30b835/manager/0.log" Jan 24 00:58:03 crc kubenswrapper[4676]: I0124 00:58:03.538068 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms_196f45b9-e656-4760-b058-e0b5c08a50d9/manager/0.log" Jan 24 00:58:03 crc kubenswrapper[4676]: I0124 00:58:03.726723 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-69647cdbc5-96fbr_07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6/operator/0.log" Jan 24 00:58:03 crc kubenswrapper[4676]: I0124 00:58:03.946832 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-kw9l8_25146f99-405a-4473-bf27-69a7195a3338/registry-server/0.log" Jan 24 00:58:04 crc kubenswrapper[4676]: I0124 00:58:04.248653 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-49nrq_53358678-d763-4b02-a157-86a57ebd0305/manager/0.log" Jan 24 00:58:04 crc kubenswrapper[4676]: I0124 00:58:04.395008 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5d646b7d76-tdqwh_d85fa79d-818f-4079-aac4-f3fa51a90e9a/manager/0.log" Jan 24 00:58:04 crc kubenswrapper[4676]: I0124 00:58:04.632459 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-h44qw_fce4c8d0-b903-4873-8c89-2f4b9dd9c05d/operator/0.log" Jan 24 00:58:04 crc kubenswrapper[4676]: I0124 00:58:04.782609 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5d5f8c4f48-md6zs_9d79c791-c851-4c4a-aa2d-d175b668b0f5/manager/0.log" Jan 24 00:58:04 crc kubenswrapper[4676]: I0124 00:58:04.852414 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5df95d5965-h8wx9_b0c8972b-31d7-40c1-bc65-1478718d41a5/manager/0.log" Jan 24 00:58:04 crc kubenswrapper[4676]: I0124 00:58:04.996936 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-4c5zl_060e1c8d-dfa6-428f-bffe-d89ac3dab8c3/manager/0.log" Jan 24 00:58:05 crc kubenswrapper[4676]: I0124 00:58:05.086631 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-gqg82_6245b73e-9fba-4ad7-bbbc-31db48c03825/manager/0.log" Jan 24 00:58:05 crc kubenswrapper[4676]: I0124 00:58:05.283333 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6d9458688d-h6hzt_ccb1ff12-bef7-4f23-b084-fae32f8202ac/manager/0.log" Jan 24 00:58:08 crc kubenswrapper[4676]: I0124 00:58:08.256266 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:58:08 crc kubenswrapper[4676]: E0124 00:58:08.256718 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.801317 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bl9lk"] Jan 24 00:58:20 crc kubenswrapper[4676]: E0124 00:58:20.802311 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e133ba60-b33b-4289-a6a7-ef3827079dbb" containerName="container-00" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.802327 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="e133ba60-b33b-4289-a6a7-ef3827079dbb" containerName="container-00" Jan 24 00:58:20 crc kubenswrapper[4676]: E0124 00:58:20.802360 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerName="registry-server" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.802368 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerName="registry-server" Jan 24 00:58:20 crc kubenswrapper[4676]: E0124 00:58:20.802397 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerName="extract-utilities" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.802407 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerName="extract-utilities" Jan 24 00:58:20 crc kubenswrapper[4676]: E0124 00:58:20.802425 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerName="extract-content" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.802433 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerName="extract-content" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.802659 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="e133ba60-b33b-4289-a6a7-ef3827079dbb" containerName="container-00" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.802677 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b3d882-428b-47c9-97e6-9c26f4334f3e" containerName="registry-server" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.804291 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.822049 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bl9lk"] Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.914034 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-utilities\") pod \"redhat-operators-bl9lk\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.914083 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k7qj\" (UniqueName: \"kubernetes.io/projected/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-kube-api-access-5k7qj\") pod \"redhat-operators-bl9lk\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:20 crc kubenswrapper[4676]: I0124 00:58:20.914192 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-catalog-content\") pod \"redhat-operators-bl9lk\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.016354 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-catalog-content\") pod \"redhat-operators-bl9lk\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.016536 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-utilities\") pod \"redhat-operators-bl9lk\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.016602 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k7qj\" (UniqueName: \"kubernetes.io/projected/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-kube-api-access-5k7qj\") pod \"redhat-operators-bl9lk\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.016841 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-catalog-content\") pod \"redhat-operators-bl9lk\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.016991 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-utilities\") pod \"redhat-operators-bl9lk\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.044429 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k7qj\" (UniqueName: \"kubernetes.io/projected/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-kube-api-access-5k7qj\") pod \"redhat-operators-bl9lk\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.140294 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.257912 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.524660 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"e910204e57fdf9a2ba78d40b8e9fd8506cf9f7184737c120d4af2680b61668e0"} Jan 24 00:58:21 crc kubenswrapper[4676]: I0124 00:58:21.622402 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bl9lk"] Jan 24 00:58:21 crc kubenswrapper[4676]: W0124 00:58:21.638511 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8a1e5fc_eb67_49a7_b2cd_2b90bee197cc.slice/crio-0dba3689f66e0cb7d0e24e7df3a5c5d6dd8739276f75ac1dc841f5d60c3282bb WatchSource:0}: Error finding container 0dba3689f66e0cb7d0e24e7df3a5c5d6dd8739276f75ac1dc841f5d60c3282bb: Status 404 returned error can't find the container with id 0dba3689f66e0cb7d0e24e7df3a5c5d6dd8739276f75ac1dc841f5d60c3282bb Jan 24 00:58:22 crc kubenswrapper[4676]: I0124 00:58:22.532493 4676 generic.go:334] "Generic (PLEG): container finished" podID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerID="688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb" exitCode=0 Jan 24 00:58:22 crc kubenswrapper[4676]: I0124 00:58:22.532707 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl9lk" event={"ID":"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc","Type":"ContainerDied","Data":"688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb"} Jan 24 00:58:22 crc kubenswrapper[4676]: I0124 00:58:22.532877 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl9lk" event={"ID":"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc","Type":"ContainerStarted","Data":"0dba3689f66e0cb7d0e24e7df3a5c5d6dd8739276f75ac1dc841f5d60c3282bb"} Jan 24 00:58:22 crc kubenswrapper[4676]: I0124 00:58:22.538419 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 00:58:24 crc kubenswrapper[4676]: I0124 00:58:24.551502 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl9lk" event={"ID":"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc","Type":"ContainerStarted","Data":"969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81"} Jan 24 00:58:27 crc kubenswrapper[4676]: I0124 00:58:27.580881 4676 generic.go:334] "Generic (PLEG): container finished" podID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerID="969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81" exitCode=0 Jan 24 00:58:27 crc kubenswrapper[4676]: I0124 00:58:27.580979 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl9lk" event={"ID":"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc","Type":"ContainerDied","Data":"969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81"} Jan 24 00:58:28 crc kubenswrapper[4676]: I0124 00:58:28.228072 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mkwj8_f53afe6a-307c-4b0d-88cb-596703f35f8a/control-plane-machine-set-operator/0.log" Jan 24 00:58:28 crc kubenswrapper[4676]: I0124 00:58:28.380703 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jmv6d_9c51c973-c370-41e8-b167-25d3b11418bf/kube-rbac-proxy/0.log" Jan 24 00:58:28 crc kubenswrapper[4676]: I0124 00:58:28.483672 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jmv6d_9c51c973-c370-41e8-b167-25d3b11418bf/machine-api-operator/0.log" Jan 24 00:58:28 crc kubenswrapper[4676]: I0124 00:58:28.592785 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl9lk" event={"ID":"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc","Type":"ContainerStarted","Data":"9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c"} Jan 24 00:58:31 crc kubenswrapper[4676]: I0124 00:58:31.141427 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:31 crc kubenswrapper[4676]: I0124 00:58:31.149163 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:32 crc kubenswrapper[4676]: I0124 00:58:32.211477 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bl9lk" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerName="registry-server" probeResult="failure" output=< Jan 24 00:58:32 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 00:58:32 crc kubenswrapper[4676]: > Jan 24 00:58:41 crc kubenswrapper[4676]: I0124 00:58:41.190162 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:41 crc kubenswrapper[4676]: I0124 00:58:41.213632 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bl9lk" podStartSLOduration=15.719740757 podStartE2EDuration="21.21361411s" podCreationTimestamp="2026-01-24 00:58:20 +0000 UTC" firstStartedPulling="2026-01-24 00:58:22.5376986 +0000 UTC m=+3286.567669601" lastFinishedPulling="2026-01-24 00:58:28.031571923 +0000 UTC m=+3292.061542954" observedRunningTime="2026-01-24 00:58:28.617358812 +0000 UTC m=+3292.647329813" watchObservedRunningTime="2026-01-24 00:58:41.21361411 +0000 UTC m=+3305.243585111" Jan 24 00:58:41 crc kubenswrapper[4676]: I0124 00:58:41.245662 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:41 crc kubenswrapper[4676]: I0124 00:58:41.423624 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bl9lk"] Jan 24 00:58:42 crc kubenswrapper[4676]: I0124 00:58:42.697819 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bl9lk" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerName="registry-server" containerID="cri-o://9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c" gracePeriod=2 Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.158611 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.311880 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-utilities\") pod \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.311955 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-catalog-content\") pod \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.312023 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k7qj\" (UniqueName: \"kubernetes.io/projected/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-kube-api-access-5k7qj\") pod \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\" (UID: \"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc\") " Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.313995 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-utilities" (OuterVolumeSpecName: "utilities") pod "b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" (UID: "b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.320490 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-kube-api-access-5k7qj" (OuterVolumeSpecName: "kube-api-access-5k7qj") pod "b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" (UID: "b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc"). InnerVolumeSpecName "kube-api-access-5k7qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.414287 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.414538 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k7qj\" (UniqueName: \"kubernetes.io/projected/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-kube-api-access-5k7qj\") on node \"crc\" DevicePath \"\"" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.442881 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" (UID: "b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.516038 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.706461 4676 generic.go:334] "Generic (PLEG): container finished" podID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerID="9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c" exitCode=0 Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.706567 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl9lk" event={"ID":"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc","Type":"ContainerDied","Data":"9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c"} Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.707483 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bl9lk" event={"ID":"b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc","Type":"ContainerDied","Data":"0dba3689f66e0cb7d0e24e7df3a5c5d6dd8739276f75ac1dc841f5d60c3282bb"} Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.706609 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bl9lk" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.707604 4676 scope.go:117] "RemoveContainer" containerID="9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.739067 4676 scope.go:117] "RemoveContainer" containerID="969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.766933 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bl9lk"] Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.777994 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bl9lk"] Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.784924 4676 scope.go:117] "RemoveContainer" containerID="688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.818811 4676 scope.go:117] "RemoveContainer" containerID="9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c" Jan 24 00:58:43 crc kubenswrapper[4676]: E0124 00:58:43.819256 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c\": container with ID starting with 9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c not found: ID does not exist" containerID="9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.819299 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c"} err="failed to get container status \"9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c\": rpc error: code = NotFound desc = could not find container \"9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c\": container with ID starting with 9d43661722191fb8d28de5b7542e25f9042c82a9ed08eb79ea6a8b5d68859f6c not found: ID does not exist" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.819328 4676 scope.go:117] "RemoveContainer" containerID="969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81" Jan 24 00:58:43 crc kubenswrapper[4676]: E0124 00:58:43.819647 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81\": container with ID starting with 969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81 not found: ID does not exist" containerID="969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.819677 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81"} err="failed to get container status \"969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81\": rpc error: code = NotFound desc = could not find container \"969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81\": container with ID starting with 969329c3edcedad71aad0fb3d89acc2cef29203d9541bdbe7d5e06de7ec62a81 not found: ID does not exist" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.819701 4676 scope.go:117] "RemoveContainer" containerID="688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb" Jan 24 00:58:43 crc kubenswrapper[4676]: E0124 00:58:43.819899 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb\": container with ID starting with 688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb not found: ID does not exist" containerID="688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb" Jan 24 00:58:43 crc kubenswrapper[4676]: I0124 00:58:43.819927 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb"} err="failed to get container status \"688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb\": rpc error: code = NotFound desc = could not find container \"688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb\": container with ID starting with 688d8fdbedb79d38d0104bb243d58508f9226f14b06d510893fa4ef3271182eb not found: ID does not exist" Jan 24 00:58:44 crc kubenswrapper[4676]: I0124 00:58:44.265233 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" path="/var/lib/kubelet/pods/b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc/volumes" Jan 24 00:58:44 crc kubenswrapper[4676]: I0124 00:58:44.392444 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-jws6v_1b51f181-5a85-4c07-b259-f67d17bf1134/cert-manager-controller/0.log" Jan 24 00:58:44 crc kubenswrapper[4676]: I0124 00:58:44.669325 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-s27fj_6db681a1-6165-4404-81a9-a189e9b30bfd/cert-manager-cainjector/0.log" Jan 24 00:58:44 crc kubenswrapper[4676]: I0124 00:58:44.694499 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-w7kb7_8f43788b-d559-4623-ae87-a820a2f23b08/cert-manager-webhook/0.log" Jan 24 00:59:00 crc kubenswrapper[4676]: I0124 00:59:00.387997 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-95988_4db34b09-85e6-435d-b991-c2513eec5d17/nmstate-console-plugin/0.log" Jan 24 00:59:00 crc kubenswrapper[4676]: I0124 00:59:00.930549 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lkqsz_6bdbc24e-ee52-41f0-9aaf-1091ac803c27/nmstate-handler/0.log" Jan 24 00:59:01 crc kubenswrapper[4676]: I0124 00:59:01.010900 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-rv2q7_0c99caac-d9eb-494c-bd04-c18dbc8a0844/kube-rbac-proxy/0.log" Jan 24 00:59:01 crc kubenswrapper[4676]: I0124 00:59:01.117542 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-rv2q7_0c99caac-d9eb-494c-bd04-c18dbc8a0844/nmstate-metrics/0.log" Jan 24 00:59:01 crc kubenswrapper[4676]: I0124 00:59:01.337248 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-cpbrv_badb3470-b60b-44ca-8d9e-52191ea016fa/nmstate-webhook/0.log" Jan 24 00:59:01 crc kubenswrapper[4676]: I0124 00:59:01.462832 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-jsbxx_efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2/nmstate-operator/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.185042 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-jqk4f_f5517856-fac7-4312-ab46-86bbd5c1282d/kube-rbac-proxy/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.232102 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-jqk4f_f5517856-fac7-4312-ab46-86bbd5c1282d/controller/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.433839 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-frr-files/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.628722 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-frr-files/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.711199 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-reloader/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.733091 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-metrics/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.739024 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-reloader/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.921940 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-frr-files/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.946846 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-metrics/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.954046 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-reloader/0.log" Jan 24 00:59:32 crc kubenswrapper[4676]: I0124 00:59:32.999848 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-metrics/0.log" Jan 24 00:59:33 crc kubenswrapper[4676]: I0124 00:59:33.291814 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-reloader/0.log" Jan 24 00:59:33 crc kubenswrapper[4676]: I0124 00:59:33.298185 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-frr-files/0.log" Jan 24 00:59:33 crc kubenswrapper[4676]: I0124 00:59:33.327871 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-metrics/0.log" Jan 24 00:59:33 crc kubenswrapper[4676]: I0124 00:59:33.339473 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/controller/0.log" Jan 24 00:59:33 crc kubenswrapper[4676]: I0124 00:59:33.539144 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/frr-metrics/0.log" Jan 24 00:59:33 crc kubenswrapper[4676]: I0124 00:59:33.540682 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/kube-rbac-proxy/0.log" Jan 24 00:59:33 crc kubenswrapper[4676]: I0124 00:59:33.683661 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/kube-rbac-proxy-frr/0.log" Jan 24 00:59:34 crc kubenswrapper[4676]: I0124 00:59:34.107237 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/reloader/0.log" Jan 24 00:59:34 crc kubenswrapper[4676]: I0124 00:59:34.197862 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-s7b59_2ce22d83-ee4f-4ad7-8882-b876d4ed52a2/frr-k8s-webhook-server/0.log" Jan 24 00:59:34 crc kubenswrapper[4676]: I0124 00:59:34.520124 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6868db74d6-qfhwb_59200cad-bd1a-472a-a1a1-adccb5211b21/manager/0.log" Jan 24 00:59:34 crc kubenswrapper[4676]: I0124 00:59:34.613009 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/frr/0.log" Jan 24 00:59:34 crc kubenswrapper[4676]: I0124 00:59:34.649371 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6dbffff8f5-cb5l5_adfe0dac-5ac5-44b8-97db-088d1ac83d34/webhook-server/0.log" Jan 24 00:59:34 crc kubenswrapper[4676]: I0124 00:59:34.767695 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zh6td_7da3019e-01de-4671-a78f-6c0d2e57fde3/kube-rbac-proxy/0.log" Jan 24 00:59:35 crc kubenswrapper[4676]: I0124 00:59:35.086796 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zh6td_7da3019e-01de-4671-a78f-6c0d2e57fde3/speaker/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.085016 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/util/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.319443 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/util/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.376204 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/pull/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.391459 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/pull/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.533576 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/util/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.543634 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/pull/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.682004 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/extract/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.768217 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/util/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.918897 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/pull/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.926013 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/pull/0.log" Jan 24 00:59:49 crc kubenswrapper[4676]: I0124 00:59:49.966519 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/util/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.084167 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/pull/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.128666 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/extract/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.146860 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/util/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.263448 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-utilities/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.490741 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-utilities/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.506660 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-content/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.527791 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-content/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.733614 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-utilities/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.767533 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-content/0.log" Jan 24 00:59:50 crc kubenswrapper[4676]: I0124 00:59:50.982194 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/registry-server/0.log" Jan 24 00:59:51 crc kubenswrapper[4676]: I0124 00:59:51.020111 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-utilities/0.log" Jan 24 00:59:51 crc kubenswrapper[4676]: I0124 00:59:51.191288 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-content/0.log" Jan 24 00:59:51 crc kubenswrapper[4676]: I0124 00:59:51.213562 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-content/0.log" Jan 24 00:59:51 crc kubenswrapper[4676]: I0124 00:59:51.216759 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-utilities/0.log" Jan 24 00:59:51 crc kubenswrapper[4676]: I0124 00:59:51.442715 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-utilities/0.log" Jan 24 00:59:51 crc kubenswrapper[4676]: I0124 00:59:51.475395 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-content/0.log" Jan 24 00:59:51 crc kubenswrapper[4676]: I0124 00:59:51.793981 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kwptt_8db31db6-2c7c-4688-89c7-328024cd7003/marketplace-operator/0.log" Jan 24 00:59:51 crc kubenswrapper[4676]: I0124 00:59:51.933784 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-utilities/0.log" Jan 24 00:59:51 crc kubenswrapper[4676]: I0124 00:59:51.953183 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/registry-server/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.115529 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-utilities/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.136412 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-content/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.164862 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-content/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.399257 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-content/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.445845 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-utilities/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.504528 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/registry-server/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.652348 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-utilities/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.817784 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-content/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.854253 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-utilities/0.log" Jan 24 00:59:52 crc kubenswrapper[4676]: I0124 00:59:52.871134 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-content/0.log" Jan 24 00:59:53 crc kubenswrapper[4676]: I0124 00:59:53.052198 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-utilities/0.log" Jan 24 00:59:53 crc kubenswrapper[4676]: I0124 00:59:53.064355 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-content/0.log" Jan 24 00:59:53 crc kubenswrapper[4676]: I0124 00:59:53.449684 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/registry-server/0.log" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.166048 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2"] Jan 24 01:00:00 crc kubenswrapper[4676]: E0124 01:00:00.167701 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerName="extract-utilities" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.167728 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerName="extract-utilities" Jan 24 01:00:00 crc kubenswrapper[4676]: E0124 01:00:00.167745 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerName="registry-server" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.167757 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerName="registry-server" Jan 24 01:00:00 crc kubenswrapper[4676]: E0124 01:00:00.167799 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerName="extract-content" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.167810 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerName="extract-content" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.168159 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8a1e5fc-eb67-49a7-b2cd-2b90bee197cc" containerName="registry-server" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.169133 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.171510 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.171565 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.177256 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2"] Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.263937 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/13892a91-56d1-49b9-9218-2c852080bfdb-kube-api-access-djwpk\") pod \"collect-profiles-29486940-s78l2\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.264642 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13892a91-56d1-49b9-9218-2c852080bfdb-config-volume\") pod \"collect-profiles-29486940-s78l2\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.264699 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13892a91-56d1-49b9-9218-2c852080bfdb-secret-volume\") pod \"collect-profiles-29486940-s78l2\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.367607 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/13892a91-56d1-49b9-9218-2c852080bfdb-kube-api-access-djwpk\") pod \"collect-profiles-29486940-s78l2\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.367677 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13892a91-56d1-49b9-9218-2c852080bfdb-config-volume\") pod \"collect-profiles-29486940-s78l2\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.367794 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13892a91-56d1-49b9-9218-2c852080bfdb-secret-volume\") pod \"collect-profiles-29486940-s78l2\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.370602 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13892a91-56d1-49b9-9218-2c852080bfdb-config-volume\") pod \"collect-profiles-29486940-s78l2\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.384157 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13892a91-56d1-49b9-9218-2c852080bfdb-secret-volume\") pod \"collect-profiles-29486940-s78l2\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.387673 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/13892a91-56d1-49b9-9218-2c852080bfdb-kube-api-access-djwpk\") pod \"collect-profiles-29486940-s78l2\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.496326 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:00 crc kubenswrapper[4676]: I0124 01:00:00.937064 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2"] Jan 24 01:00:00 crc kubenswrapper[4676]: W0124 01:00:00.947634 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13892a91_56d1_49b9_9218_2c852080bfdb.slice/crio-d92f79a693b27138ac891230baed68cfe599a074ee62367fdafee3f9f2dae67c WatchSource:0}: Error finding container d92f79a693b27138ac891230baed68cfe599a074ee62367fdafee3f9f2dae67c: Status 404 returned error can't find the container with id d92f79a693b27138ac891230baed68cfe599a074ee62367fdafee3f9f2dae67c Jan 24 01:00:01 crc kubenswrapper[4676]: I0124 01:00:01.340687 4676 generic.go:334] "Generic (PLEG): container finished" podID="13892a91-56d1-49b9-9218-2c852080bfdb" containerID="a1167e023def9f55810d9e752100b280bff41604a70b8d00cf03ee7930fce39d" exitCode=0 Jan 24 01:00:01 crc kubenswrapper[4676]: I0124 01:00:01.340758 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" event={"ID":"13892a91-56d1-49b9-9218-2c852080bfdb","Type":"ContainerDied","Data":"a1167e023def9f55810d9e752100b280bff41604a70b8d00cf03ee7930fce39d"} Jan 24 01:00:01 crc kubenswrapper[4676]: I0124 01:00:01.340977 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" event={"ID":"13892a91-56d1-49b9-9218-2c852080bfdb","Type":"ContainerStarted","Data":"d92f79a693b27138ac891230baed68cfe599a074ee62367fdafee3f9f2dae67c"} Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.718221 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.810670 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13892a91-56d1-49b9-9218-2c852080bfdb-secret-volume\") pod \"13892a91-56d1-49b9-9218-2c852080bfdb\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.810726 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13892a91-56d1-49b9-9218-2c852080bfdb-config-volume\") pod \"13892a91-56d1-49b9-9218-2c852080bfdb\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.810894 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/13892a91-56d1-49b9-9218-2c852080bfdb-kube-api-access-djwpk\") pod \"13892a91-56d1-49b9-9218-2c852080bfdb\" (UID: \"13892a91-56d1-49b9-9218-2c852080bfdb\") " Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.811961 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13892a91-56d1-49b9-9218-2c852080bfdb-config-volume" (OuterVolumeSpecName: "config-volume") pod "13892a91-56d1-49b9-9218-2c852080bfdb" (UID: "13892a91-56d1-49b9-9218-2c852080bfdb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.817986 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13892a91-56d1-49b9-9218-2c852080bfdb-kube-api-access-djwpk" (OuterVolumeSpecName: "kube-api-access-djwpk") pod "13892a91-56d1-49b9-9218-2c852080bfdb" (UID: "13892a91-56d1-49b9-9218-2c852080bfdb"). InnerVolumeSpecName "kube-api-access-djwpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.818493 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13892a91-56d1-49b9-9218-2c852080bfdb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "13892a91-56d1-49b9-9218-2c852080bfdb" (UID: "13892a91-56d1-49b9-9218-2c852080bfdb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.913237 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djwpk\" (UniqueName: \"kubernetes.io/projected/13892a91-56d1-49b9-9218-2c852080bfdb-kube-api-access-djwpk\") on node \"crc\" DevicePath \"\"" Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.913275 4676 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/13892a91-56d1-49b9-9218-2c852080bfdb-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 24 01:00:02 crc kubenswrapper[4676]: I0124 01:00:02.913287 4676 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/13892a91-56d1-49b9-9218-2c852080bfdb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 24 01:00:03 crc kubenswrapper[4676]: I0124 01:00:03.360511 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" event={"ID":"13892a91-56d1-49b9-9218-2c852080bfdb","Type":"ContainerDied","Data":"d92f79a693b27138ac891230baed68cfe599a074ee62367fdafee3f9f2dae67c"} Jan 24 01:00:03 crc kubenswrapper[4676]: I0124 01:00:03.360550 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d92f79a693b27138ac891230baed68cfe599a074ee62367fdafee3f9f2dae67c" Jan 24 01:00:03 crc kubenswrapper[4676]: I0124 01:00:03.360598 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29486940-s78l2" Jan 24 01:00:03 crc kubenswrapper[4676]: I0124 01:00:03.813334 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p"] Jan 24 01:00:03 crc kubenswrapper[4676]: I0124 01:00:03.825255 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29486895-5v67p"] Jan 24 01:00:04 crc kubenswrapper[4676]: I0124 01:00:04.269216 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40ef8e58-50e2-4cc5-b92c-35710605e5b1" path="/var/lib/kubelet/pods/40ef8e58-50e2-4cc5-b92c-35710605e5b1/volumes" Jan 24 01:00:25 crc kubenswrapper[4676]: E0124 01:00:25.369851 4676 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.27:59720->38.102.83.27:34811: write tcp 38.102.83.27:59720->38.102.83.27:34811: write: broken pipe Jan 24 01:00:39 crc kubenswrapper[4676]: I0124 01:00:39.364726 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 01:00:39 crc kubenswrapper[4676]: I0124 01:00:39.365210 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 01:00:43 crc kubenswrapper[4676]: I0124 01:00:43.272788 4676 scope.go:117] "RemoveContainer" containerID="1f9529fca33e87b8d8a2cf842fe194d301e46a27673eb8a957977d0bcb8f4f8d" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.157550 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29486941-6wtvm"] Jan 24 01:01:00 crc kubenswrapper[4676]: E0124 01:01:00.158777 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13892a91-56d1-49b9-9218-2c852080bfdb" containerName="collect-profiles" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.158801 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="13892a91-56d1-49b9-9218-2c852080bfdb" containerName="collect-profiles" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.159189 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="13892a91-56d1-49b9-9218-2c852080bfdb" containerName="collect-profiles" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.160266 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.174344 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29486941-6wtvm"] Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.247570 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdrql\" (UniqueName: \"kubernetes.io/projected/483961de-208d-4593-a6c8-ecee687b7c06-kube-api-access-mdrql\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.247752 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-config-data\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.247779 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-fernet-keys\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.247813 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-combined-ca-bundle\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.350428 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-config-data\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.350483 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-fernet-keys\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.350519 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-combined-ca-bundle\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.350619 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdrql\" (UniqueName: \"kubernetes.io/projected/483961de-208d-4593-a6c8-ecee687b7c06-kube-api-access-mdrql\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.358792 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-combined-ca-bundle\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.362684 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-fernet-keys\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.370201 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-config-data\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.392797 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdrql\" (UniqueName: \"kubernetes.io/projected/483961de-208d-4593-a6c8-ecee687b7c06-kube-api-access-mdrql\") pod \"keystone-cron-29486941-6wtvm\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.497082 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:00 crc kubenswrapper[4676]: I0124 01:01:00.960583 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29486941-6wtvm"] Jan 24 01:01:01 crc kubenswrapper[4676]: I0124 01:01:01.964953 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29486941-6wtvm" event={"ID":"483961de-208d-4593-a6c8-ecee687b7c06","Type":"ContainerStarted","Data":"f64c0d4c1559404ae786b4129234e03dc73707e5aea1115b52503a990f8b5b98"} Jan 24 01:01:01 crc kubenswrapper[4676]: I0124 01:01:01.965303 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29486941-6wtvm" event={"ID":"483961de-208d-4593-a6c8-ecee687b7c06","Type":"ContainerStarted","Data":"2a967c567bd8fd03d2dede0c91a2bbb28ebdd22ba30ad8ef3f097c16c69a4a6e"} Jan 24 01:01:03 crc kubenswrapper[4676]: I0124 01:01:03.983151 4676 generic.go:334] "Generic (PLEG): container finished" podID="483961de-208d-4593-a6c8-ecee687b7c06" containerID="f64c0d4c1559404ae786b4129234e03dc73707e5aea1115b52503a990f8b5b98" exitCode=0 Jan 24 01:01:03 crc kubenswrapper[4676]: I0124 01:01:03.983259 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29486941-6wtvm" event={"ID":"483961de-208d-4593-a6c8-ecee687b7c06","Type":"ContainerDied","Data":"f64c0d4c1559404ae786b4129234e03dc73707e5aea1115b52503a990f8b5b98"} Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.342026 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.453629 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdrql\" (UniqueName: \"kubernetes.io/projected/483961de-208d-4593-a6c8-ecee687b7c06-kube-api-access-mdrql\") pod \"483961de-208d-4593-a6c8-ecee687b7c06\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.453994 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-fernet-keys\") pod \"483961de-208d-4593-a6c8-ecee687b7c06\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.454057 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-combined-ca-bundle\") pod \"483961de-208d-4593-a6c8-ecee687b7c06\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.454203 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-config-data\") pod \"483961de-208d-4593-a6c8-ecee687b7c06\" (UID: \"483961de-208d-4593-a6c8-ecee687b7c06\") " Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.462838 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "483961de-208d-4593-a6c8-ecee687b7c06" (UID: "483961de-208d-4593-a6c8-ecee687b7c06"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.464581 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/483961de-208d-4593-a6c8-ecee687b7c06-kube-api-access-mdrql" (OuterVolumeSpecName: "kube-api-access-mdrql") pod "483961de-208d-4593-a6c8-ecee687b7c06" (UID: "483961de-208d-4593-a6c8-ecee687b7c06"). InnerVolumeSpecName "kube-api-access-mdrql". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.490635 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "483961de-208d-4593-a6c8-ecee687b7c06" (UID: "483961de-208d-4593-a6c8-ecee687b7c06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.523683 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-config-data" (OuterVolumeSpecName: "config-data") pod "483961de-208d-4593-a6c8-ecee687b7c06" (UID: "483961de-208d-4593-a6c8-ecee687b7c06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.557537 4676 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-config-data\") on node \"crc\" DevicePath \"\"" Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.557576 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdrql\" (UniqueName: \"kubernetes.io/projected/483961de-208d-4593-a6c8-ecee687b7c06-kube-api-access-mdrql\") on node \"crc\" DevicePath \"\"" Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.557587 4676 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 24 01:01:05 crc kubenswrapper[4676]: I0124 01:01:05.557596 4676 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483961de-208d-4593-a6c8-ecee687b7c06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 24 01:01:06 crc kubenswrapper[4676]: I0124 01:01:06.001896 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29486941-6wtvm" event={"ID":"483961de-208d-4593-a6c8-ecee687b7c06","Type":"ContainerDied","Data":"2a967c567bd8fd03d2dede0c91a2bbb28ebdd22ba30ad8ef3f097c16c69a4a6e"} Jan 24 01:01:06 crc kubenswrapper[4676]: I0124 01:01:06.001933 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a967c567bd8fd03d2dede0c91a2bbb28ebdd22ba30ad8ef3f097c16c69a4a6e" Jan 24 01:01:06 crc kubenswrapper[4676]: I0124 01:01:06.001982 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29486941-6wtvm" Jan 24 01:01:09 crc kubenswrapper[4676]: I0124 01:01:09.364291 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 01:01:09 crc kubenswrapper[4676]: I0124 01:01:09.365024 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 01:01:39 crc kubenswrapper[4676]: I0124 01:01:39.363879 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 01:01:39 crc kubenswrapper[4676]: I0124 01:01:39.364527 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 01:01:39 crc kubenswrapper[4676]: I0124 01:01:39.364590 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 01:01:39 crc kubenswrapper[4676]: I0124 01:01:39.365588 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e910204e57fdf9a2ba78d40b8e9fd8506cf9f7184737c120d4af2680b61668e0"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 01:01:39 crc kubenswrapper[4676]: I0124 01:01:39.365684 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://e910204e57fdf9a2ba78d40b8e9fd8506cf9f7184737c120d4af2680b61668e0" gracePeriod=600 Jan 24 01:01:40 crc kubenswrapper[4676]: I0124 01:01:40.440901 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="e910204e57fdf9a2ba78d40b8e9fd8506cf9f7184737c120d4af2680b61668e0" exitCode=0 Jan 24 01:01:40 crc kubenswrapper[4676]: I0124 01:01:40.440963 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"e910204e57fdf9a2ba78d40b8e9fd8506cf9f7184737c120d4af2680b61668e0"} Jan 24 01:01:40 crc kubenswrapper[4676]: I0124 01:01:40.441405 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9"} Jan 24 01:01:40 crc kubenswrapper[4676]: I0124 01:01:40.441442 4676 scope.go:117] "RemoveContainer" containerID="3028d9f3374dc67ac159b532c9f99a34d8cb2b7f6bd88b8c4da01f1590e10398" Jan 24 01:01:41 crc kubenswrapper[4676]: I0124 01:01:41.458637 4676 generic.go:334] "Generic (PLEG): container finished" podID="a646626a-4429-4d8a-9f75-7b8bfade90ec" containerID="e762fbd56fb888c091ce2f8ff2b6ad8a52059af7b071066cce8729c1ec864419" exitCode=0 Jan 24 01:01:41 crc kubenswrapper[4676]: I0124 01:01:41.458724 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-57xmz/must-gather-hv4gr" event={"ID":"a646626a-4429-4d8a-9f75-7b8bfade90ec","Type":"ContainerDied","Data":"e762fbd56fb888c091ce2f8ff2b6ad8a52059af7b071066cce8729c1ec864419"} Jan 24 01:01:41 crc kubenswrapper[4676]: I0124 01:01:41.460304 4676 scope.go:117] "RemoveContainer" containerID="e762fbd56fb888c091ce2f8ff2b6ad8a52059af7b071066cce8729c1ec864419" Jan 24 01:01:41 crc kubenswrapper[4676]: I0124 01:01:41.872143 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-57xmz_must-gather-hv4gr_a646626a-4429-4d8a-9f75-7b8bfade90ec/gather/0.log" Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.296804 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-57xmz/must-gather-hv4gr"] Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.297552 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-57xmz/must-gather-hv4gr" podUID="a646626a-4429-4d8a-9f75-7b8bfade90ec" containerName="copy" containerID="cri-o://5282e9521531b7fbcb12f5cd407d932cd2321b1dd6ad9e7d8b88355c7f251d0b" gracePeriod=2 Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.336190 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-57xmz/must-gather-hv4gr"] Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.563931 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-57xmz_must-gather-hv4gr_a646626a-4429-4d8a-9f75-7b8bfade90ec/copy/0.log" Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.564570 4676 generic.go:334] "Generic (PLEG): container finished" podID="a646626a-4429-4d8a-9f75-7b8bfade90ec" containerID="5282e9521531b7fbcb12f5cd407d932cd2321b1dd6ad9e7d8b88355c7f251d0b" exitCode=143 Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.825186 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-57xmz_must-gather-hv4gr_a646626a-4429-4d8a-9f75-7b8bfade90ec/copy/0.log" Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.825819 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.907238 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbztc\" (UniqueName: \"kubernetes.io/projected/a646626a-4429-4d8a-9f75-7b8bfade90ec-kube-api-access-jbztc\") pod \"a646626a-4429-4d8a-9f75-7b8bfade90ec\" (UID: \"a646626a-4429-4d8a-9f75-7b8bfade90ec\") " Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.907437 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a646626a-4429-4d8a-9f75-7b8bfade90ec-must-gather-output\") pod \"a646626a-4429-4d8a-9f75-7b8bfade90ec\" (UID: \"a646626a-4429-4d8a-9f75-7b8bfade90ec\") " Jan 24 01:01:50 crc kubenswrapper[4676]: I0124 01:01:50.913077 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a646626a-4429-4d8a-9f75-7b8bfade90ec-kube-api-access-jbztc" (OuterVolumeSpecName: "kube-api-access-jbztc") pod "a646626a-4429-4d8a-9f75-7b8bfade90ec" (UID: "a646626a-4429-4d8a-9f75-7b8bfade90ec"). InnerVolumeSpecName "kube-api-access-jbztc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:01:51 crc kubenswrapper[4676]: I0124 01:01:51.009031 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbztc\" (UniqueName: \"kubernetes.io/projected/a646626a-4429-4d8a-9f75-7b8bfade90ec-kube-api-access-jbztc\") on node \"crc\" DevicePath \"\"" Jan 24 01:01:51 crc kubenswrapper[4676]: I0124 01:01:51.056722 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a646626a-4429-4d8a-9f75-7b8bfade90ec-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a646626a-4429-4d8a-9f75-7b8bfade90ec" (UID: "a646626a-4429-4d8a-9f75-7b8bfade90ec"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:01:51 crc kubenswrapper[4676]: I0124 01:01:51.111086 4676 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a646626a-4429-4d8a-9f75-7b8bfade90ec-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 24 01:01:51 crc kubenswrapper[4676]: I0124 01:01:51.578149 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-57xmz_must-gather-hv4gr_a646626a-4429-4d8a-9f75-7b8bfade90ec/copy/0.log" Jan 24 01:01:51 crc kubenswrapper[4676]: I0124 01:01:51.579059 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-57xmz/must-gather-hv4gr" Jan 24 01:01:51 crc kubenswrapper[4676]: I0124 01:01:51.579052 4676 scope.go:117] "RemoveContainer" containerID="5282e9521531b7fbcb12f5cd407d932cd2321b1dd6ad9e7d8b88355c7f251d0b" Jan 24 01:01:51 crc kubenswrapper[4676]: I0124 01:01:51.608269 4676 scope.go:117] "RemoveContainer" containerID="e762fbd56fb888c091ce2f8ff2b6ad8a52059af7b071066cce8729c1ec864419" Jan 24 01:01:52 crc kubenswrapper[4676]: I0124 01:01:52.274054 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a646626a-4429-4d8a-9f75-7b8bfade90ec" path="/var/lib/kubelet/pods/a646626a-4429-4d8a-9f75-7b8bfade90ec/volumes" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.345512 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dg4k2"] Jan 24 01:02:33 crc kubenswrapper[4676]: E0124 01:02:33.346410 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483961de-208d-4593-a6c8-ecee687b7c06" containerName="keystone-cron" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.346423 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="483961de-208d-4593-a6c8-ecee687b7c06" containerName="keystone-cron" Jan 24 01:02:33 crc kubenswrapper[4676]: E0124 01:02:33.346437 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a646626a-4429-4d8a-9f75-7b8bfade90ec" containerName="copy" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.346442 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a646626a-4429-4d8a-9f75-7b8bfade90ec" containerName="copy" Jan 24 01:02:33 crc kubenswrapper[4676]: E0124 01:02:33.346450 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a646626a-4429-4d8a-9f75-7b8bfade90ec" containerName="gather" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.346456 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="a646626a-4429-4d8a-9f75-7b8bfade90ec" containerName="gather" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.346626 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="a646626a-4429-4d8a-9f75-7b8bfade90ec" containerName="gather" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.346641 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="483961de-208d-4593-a6c8-ecee687b7c06" containerName="keystone-cron" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.346657 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="a646626a-4429-4d8a-9f75-7b8bfade90ec" containerName="copy" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.348692 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.354821 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg4k2"] Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.460539 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-catalog-content\") pod \"redhat-marketplace-dg4k2\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.460600 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fptfh\" (UniqueName: \"kubernetes.io/projected/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-kube-api-access-fptfh\") pod \"redhat-marketplace-dg4k2\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.460640 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-utilities\") pod \"redhat-marketplace-dg4k2\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.562522 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-catalog-content\") pod \"redhat-marketplace-dg4k2\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.562575 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fptfh\" (UniqueName: \"kubernetes.io/projected/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-kube-api-access-fptfh\") pod \"redhat-marketplace-dg4k2\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.562616 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-utilities\") pod \"redhat-marketplace-dg4k2\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.563027 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-utilities\") pod \"redhat-marketplace-dg4k2\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.563214 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-catalog-content\") pod \"redhat-marketplace-dg4k2\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.591113 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fptfh\" (UniqueName: \"kubernetes.io/projected/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-kube-api-access-fptfh\") pod \"redhat-marketplace-dg4k2\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:33 crc kubenswrapper[4676]: I0124 01:02:33.666474 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:34 crc kubenswrapper[4676]: I0124 01:02:34.146361 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg4k2"] Jan 24 01:02:35 crc kubenswrapper[4676]: I0124 01:02:35.022362 4676 generic.go:334] "Generic (PLEG): container finished" podID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerID="1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36" exitCode=0 Jan 24 01:02:35 crc kubenswrapper[4676]: I0124 01:02:35.022529 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg4k2" event={"ID":"04f6e0b9-60c2-4233-9ae1-466f0d9856f7","Type":"ContainerDied","Data":"1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36"} Jan 24 01:02:35 crc kubenswrapper[4676]: I0124 01:02:35.023024 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg4k2" event={"ID":"04f6e0b9-60c2-4233-9ae1-466f0d9856f7","Type":"ContainerStarted","Data":"26d6623bd7a9f3a00f3aeb656c92a96d75069eab3735f0a291f1ae75f99a99cc"} Jan 24 01:02:36 crc kubenswrapper[4676]: I0124 01:02:36.036397 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg4k2" event={"ID":"04f6e0b9-60c2-4233-9ae1-466f0d9856f7","Type":"ContainerStarted","Data":"8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155"} Jan 24 01:02:37 crc kubenswrapper[4676]: I0124 01:02:37.049682 4676 generic.go:334] "Generic (PLEG): container finished" podID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerID="8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155" exitCode=0 Jan 24 01:02:37 crc kubenswrapper[4676]: I0124 01:02:37.049783 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg4k2" event={"ID":"04f6e0b9-60c2-4233-9ae1-466f0d9856f7","Type":"ContainerDied","Data":"8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155"} Jan 24 01:02:38 crc kubenswrapper[4676]: I0124 01:02:38.073164 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg4k2" event={"ID":"04f6e0b9-60c2-4233-9ae1-466f0d9856f7","Type":"ContainerStarted","Data":"c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93"} Jan 24 01:02:38 crc kubenswrapper[4676]: I0124 01:02:38.103531 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dg4k2" podStartSLOduration=2.604491114 podStartE2EDuration="5.103515592s" podCreationTimestamp="2026-01-24 01:02:33 +0000 UTC" firstStartedPulling="2026-01-24 01:02:35.024477618 +0000 UTC m=+3539.054448659" lastFinishedPulling="2026-01-24 01:02:37.523502126 +0000 UTC m=+3541.553473137" observedRunningTime="2026-01-24 01:02:38.099095085 +0000 UTC m=+3542.129066076" watchObservedRunningTime="2026-01-24 01:02:38.103515592 +0000 UTC m=+3542.133486593" Jan 24 01:02:43 crc kubenswrapper[4676]: I0124 01:02:43.412159 4676 scope.go:117] "RemoveContainer" containerID="fb528b96bb0ea9229c8a0baf4db69e4aad9f43be8642c6a3de48a10855ae6586" Jan 24 01:02:43 crc kubenswrapper[4676]: I0124 01:02:43.667704 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:43 crc kubenswrapper[4676]: I0124 01:02:43.667760 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:43 crc kubenswrapper[4676]: I0124 01:02:43.720984 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:44 crc kubenswrapper[4676]: I0124 01:02:44.179553 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:45 crc kubenswrapper[4676]: I0124 01:02:45.135578 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg4k2"] Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.153228 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dg4k2" podUID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerName="registry-server" containerID="cri-o://c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93" gracePeriod=2 Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.616624 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.647962 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fptfh\" (UniqueName: \"kubernetes.io/projected/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-kube-api-access-fptfh\") pod \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.648176 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-catalog-content\") pod \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.648298 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-utilities\") pod \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\" (UID: \"04f6e0b9-60c2-4233-9ae1-466f0d9856f7\") " Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.651701 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-utilities" (OuterVolumeSpecName: "utilities") pod "04f6e0b9-60c2-4233-9ae1-466f0d9856f7" (UID: "04f6e0b9-60c2-4233-9ae1-466f0d9856f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.677570 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-kube-api-access-fptfh" (OuterVolumeSpecName: "kube-api-access-fptfh") pod "04f6e0b9-60c2-4233-9ae1-466f0d9856f7" (UID: "04f6e0b9-60c2-4233-9ae1-466f0d9856f7"). InnerVolumeSpecName "kube-api-access-fptfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.750420 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.750451 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fptfh\" (UniqueName: \"kubernetes.io/projected/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-kube-api-access-fptfh\") on node \"crc\" DevicePath \"\"" Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.806221 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "04f6e0b9-60c2-4233-9ae1-466f0d9856f7" (UID: "04f6e0b9-60c2-4233-9ae1-466f0d9856f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:02:46 crc kubenswrapper[4676]: I0124 01:02:46.853573 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04f6e0b9-60c2-4233-9ae1-466f0d9856f7-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.163990 4676 generic.go:334] "Generic (PLEG): container finished" podID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerID="c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93" exitCode=0 Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.164068 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dg4k2" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.164092 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg4k2" event={"ID":"04f6e0b9-60c2-4233-9ae1-466f0d9856f7","Type":"ContainerDied","Data":"c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93"} Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.165488 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dg4k2" event={"ID":"04f6e0b9-60c2-4233-9ae1-466f0d9856f7","Type":"ContainerDied","Data":"26d6623bd7a9f3a00f3aeb656c92a96d75069eab3735f0a291f1ae75f99a99cc"} Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.165594 4676 scope.go:117] "RemoveContainer" containerID="c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.210223 4676 scope.go:117] "RemoveContainer" containerID="8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.213259 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg4k2"] Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.226831 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dg4k2"] Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.236156 4676 scope.go:117] "RemoveContainer" containerID="1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.278890 4676 scope.go:117] "RemoveContainer" containerID="c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93" Jan 24 01:02:47 crc kubenswrapper[4676]: E0124 01:02:47.279342 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93\": container with ID starting with c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93 not found: ID does not exist" containerID="c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.279504 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93"} err="failed to get container status \"c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93\": rpc error: code = NotFound desc = could not find container \"c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93\": container with ID starting with c92d3b1e742077bc21525575f682ff78f7de491d3f8231de5ae308c3c85f8e93 not found: ID does not exist" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.279549 4676 scope.go:117] "RemoveContainer" containerID="8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155" Jan 24 01:02:47 crc kubenswrapper[4676]: E0124 01:02:47.280845 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155\": container with ID starting with 8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155 not found: ID does not exist" containerID="8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.281312 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155"} err="failed to get container status \"8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155\": rpc error: code = NotFound desc = could not find container \"8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155\": container with ID starting with 8d31fcde81cecd6db5fb594b425698747416b475844d88a2e2664f5bd76c2155 not found: ID does not exist" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.281715 4676 scope.go:117] "RemoveContainer" containerID="1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36" Jan 24 01:02:47 crc kubenswrapper[4676]: E0124 01:02:47.282728 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36\": container with ID starting with 1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36 not found: ID does not exist" containerID="1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36" Jan 24 01:02:47 crc kubenswrapper[4676]: I0124 01:02:47.282816 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36"} err="failed to get container status \"1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36\": rpc error: code = NotFound desc = could not find container \"1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36\": container with ID starting with 1d59967b92bf2bb19b28ac9354c7eaf360e79496f4387e20edf7d4d9289aee36 not found: ID does not exist" Jan 24 01:02:48 crc kubenswrapper[4676]: I0124 01:02:48.283712 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" path="/var/lib/kubelet/pods/04f6e0b9-60c2-4233-9ae1-466f0d9856f7/volumes" Jan 24 01:03:39 crc kubenswrapper[4676]: I0124 01:03:39.364234 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 01:03:39 crc kubenswrapper[4676]: I0124 01:03:39.364927 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 01:04:09 crc kubenswrapper[4676]: I0124 01:04:09.364030 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 01:04:09 crc kubenswrapper[4676]: I0124 01:04:09.364912 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.523370 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8qqfp"] Jan 24 01:04:14 crc kubenswrapper[4676]: E0124 01:04:14.524615 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerName="registry-server" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.524639 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerName="registry-server" Jan 24 01:04:14 crc kubenswrapper[4676]: E0124 01:04:14.524674 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerName="extract-utilities" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.524686 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerName="extract-utilities" Jan 24 01:04:14 crc kubenswrapper[4676]: E0124 01:04:14.524711 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerName="extract-content" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.524722 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerName="extract-content" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.525027 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="04f6e0b9-60c2-4233-9ae1-466f0d9856f7" containerName="registry-server" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.526866 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.539560 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8qqfp"] Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.657862 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-catalog-content\") pod \"certified-operators-8qqfp\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.657984 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-utilities\") pod \"certified-operators-8qqfp\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.658019 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr5km\" (UniqueName: \"kubernetes.io/projected/45ec6f03-36a9-4179-aee2-d3ba73569525-kube-api-access-zr5km\") pod \"certified-operators-8qqfp\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.759327 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-catalog-content\") pod \"certified-operators-8qqfp\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.759511 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-utilities\") pod \"certified-operators-8qqfp\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.759562 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr5km\" (UniqueName: \"kubernetes.io/projected/45ec6f03-36a9-4179-aee2-d3ba73569525-kube-api-access-zr5km\") pod \"certified-operators-8qqfp\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.760483 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-catalog-content\") pod \"certified-operators-8qqfp\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.760766 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-utilities\") pod \"certified-operators-8qqfp\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.801009 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr5km\" (UniqueName: \"kubernetes.io/projected/45ec6f03-36a9-4179-aee2-d3ba73569525-kube-api-access-zr5km\") pod \"certified-operators-8qqfp\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:14 crc kubenswrapper[4676]: I0124 01:04:14.889430 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:15 crc kubenswrapper[4676]: I0124 01:04:15.648050 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8qqfp"] Jan 24 01:04:16 crc kubenswrapper[4676]: I0124 01:04:16.182934 4676 generic.go:334] "Generic (PLEG): container finished" podID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerID="836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721" exitCode=0 Jan 24 01:04:16 crc kubenswrapper[4676]: I0124 01:04:16.183019 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qqfp" event={"ID":"45ec6f03-36a9-4179-aee2-d3ba73569525","Type":"ContainerDied","Data":"836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721"} Jan 24 01:04:16 crc kubenswrapper[4676]: I0124 01:04:16.183092 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qqfp" event={"ID":"45ec6f03-36a9-4179-aee2-d3ba73569525","Type":"ContainerStarted","Data":"b754be22bda55b2d969d9d4307b2e0808737203d0f55c6fb5a2728ac67d3f9dd"} Jan 24 01:04:16 crc kubenswrapper[4676]: I0124 01:04:16.184780 4676 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 24 01:04:17 crc kubenswrapper[4676]: I0124 01:04:17.193743 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qqfp" event={"ID":"45ec6f03-36a9-4179-aee2-d3ba73569525","Type":"ContainerStarted","Data":"dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36"} Jan 24 01:04:18 crc kubenswrapper[4676]: I0124 01:04:18.207555 4676 generic.go:334] "Generic (PLEG): container finished" podID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerID="dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36" exitCode=0 Jan 24 01:04:18 crc kubenswrapper[4676]: I0124 01:04:18.207907 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qqfp" event={"ID":"45ec6f03-36a9-4179-aee2-d3ba73569525","Type":"ContainerDied","Data":"dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36"} Jan 24 01:04:19 crc kubenswrapper[4676]: I0124 01:04:19.226066 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qqfp" event={"ID":"45ec6f03-36a9-4179-aee2-d3ba73569525","Type":"ContainerStarted","Data":"1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3"} Jan 24 01:04:19 crc kubenswrapper[4676]: I0124 01:04:19.262010 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8qqfp" podStartSLOduration=2.852433186 podStartE2EDuration="5.261992591s" podCreationTimestamp="2026-01-24 01:04:14 +0000 UTC" firstStartedPulling="2026-01-24 01:04:16.184532326 +0000 UTC m=+3640.214503327" lastFinishedPulling="2026-01-24 01:04:18.594091691 +0000 UTC m=+3642.624062732" observedRunningTime="2026-01-24 01:04:19.255504321 +0000 UTC m=+3643.285475322" watchObservedRunningTime="2026-01-24 01:04:19.261992591 +0000 UTC m=+3643.291963592" Jan 24 01:04:24 crc kubenswrapper[4676]: I0124 01:04:24.890523 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:24 crc kubenswrapper[4676]: I0124 01:04:24.891183 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:25 crc kubenswrapper[4676]: I0124 01:04:25.019122 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:25 crc kubenswrapper[4676]: I0124 01:04:25.326887 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:25 crc kubenswrapper[4676]: I0124 01:04:25.387069 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8qqfp"] Jan 24 01:04:27 crc kubenswrapper[4676]: I0124 01:04:27.313233 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8qqfp" podUID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerName="registry-server" containerID="cri-o://1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3" gracePeriod=2 Jan 24 01:04:27 crc kubenswrapper[4676]: I0124 01:04:27.946248 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.062749 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-utilities\") pod \"45ec6f03-36a9-4179-aee2-d3ba73569525\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.062887 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-catalog-content\") pod \"45ec6f03-36a9-4179-aee2-d3ba73569525\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.062928 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr5km\" (UniqueName: \"kubernetes.io/projected/45ec6f03-36a9-4179-aee2-d3ba73569525-kube-api-access-zr5km\") pod \"45ec6f03-36a9-4179-aee2-d3ba73569525\" (UID: \"45ec6f03-36a9-4179-aee2-d3ba73569525\") " Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.064041 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-utilities" (OuterVolumeSpecName: "utilities") pod "45ec6f03-36a9-4179-aee2-d3ba73569525" (UID: "45ec6f03-36a9-4179-aee2-d3ba73569525"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.071648 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45ec6f03-36a9-4179-aee2-d3ba73569525-kube-api-access-zr5km" (OuterVolumeSpecName: "kube-api-access-zr5km") pod "45ec6f03-36a9-4179-aee2-d3ba73569525" (UID: "45ec6f03-36a9-4179-aee2-d3ba73569525"). InnerVolumeSpecName "kube-api-access-zr5km". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.111522 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45ec6f03-36a9-4179-aee2-d3ba73569525" (UID: "45ec6f03-36a9-4179-aee2-d3ba73569525"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.165305 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.165341 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45ec6f03-36a9-4179-aee2-d3ba73569525-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.165354 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr5km\" (UniqueName: \"kubernetes.io/projected/45ec6f03-36a9-4179-aee2-d3ba73569525-kube-api-access-zr5km\") on node \"crc\" DevicePath \"\"" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.324029 4676 generic.go:334] "Generic (PLEG): container finished" podID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerID="1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3" exitCode=0 Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.324065 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qqfp" event={"ID":"45ec6f03-36a9-4179-aee2-d3ba73569525","Type":"ContainerDied","Data":"1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3"} Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.324116 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8qqfp" event={"ID":"45ec6f03-36a9-4179-aee2-d3ba73569525","Type":"ContainerDied","Data":"b754be22bda55b2d969d9d4307b2e0808737203d0f55c6fb5a2728ac67d3f9dd"} Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.324139 4676 scope.go:117] "RemoveContainer" containerID="1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.324156 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8qqfp" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.347141 4676 scope.go:117] "RemoveContainer" containerID="dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.355944 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8qqfp"] Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.379103 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8qqfp"] Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.388244 4676 scope.go:117] "RemoveContainer" containerID="836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.444027 4676 scope.go:117] "RemoveContainer" containerID="1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3" Jan 24 01:04:28 crc kubenswrapper[4676]: E0124 01:04:28.444608 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3\": container with ID starting with 1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3 not found: ID does not exist" containerID="1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.445366 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3"} err="failed to get container status \"1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3\": rpc error: code = NotFound desc = could not find container \"1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3\": container with ID starting with 1c023458df9b94edf9a49905a4b340c8073ab2d77f915f786ec88a9ad9a156e3 not found: ID does not exist" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.445678 4676 scope.go:117] "RemoveContainer" containerID="dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36" Jan 24 01:04:28 crc kubenswrapper[4676]: E0124 01:04:28.446685 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36\": container with ID starting with dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36 not found: ID does not exist" containerID="dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.446803 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36"} err="failed to get container status \"dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36\": rpc error: code = NotFound desc = could not find container \"dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36\": container with ID starting with dfd9d9856f7a8afa17e0f5cfe7c12b474543a7274fba591c7fb28969982c3c36 not found: ID does not exist" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.446909 4676 scope.go:117] "RemoveContainer" containerID="836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721" Jan 24 01:04:28 crc kubenswrapper[4676]: E0124 01:04:28.447324 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721\": container with ID starting with 836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721 not found: ID does not exist" containerID="836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721" Jan 24 01:04:28 crc kubenswrapper[4676]: I0124 01:04:28.447359 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721"} err="failed to get container status \"836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721\": rpc error: code = NotFound desc = could not find container \"836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721\": container with ID starting with 836cd904b9af0c0d701a50e6dc3fae19ab1fc0c795615cfe51e9b14e16cea721 not found: ID does not exist" Jan 24 01:04:30 crc kubenswrapper[4676]: I0124 01:04:30.267420 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45ec6f03-36a9-4179-aee2-d3ba73569525" path="/var/lib/kubelet/pods/45ec6f03-36a9-4179-aee2-d3ba73569525/volumes" Jan 24 01:04:39 crc kubenswrapper[4676]: I0124 01:04:39.364453 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 01:04:39 crc kubenswrapper[4676]: I0124 01:04:39.365055 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 01:04:39 crc kubenswrapper[4676]: I0124 01:04:39.365120 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 01:04:39 crc kubenswrapper[4676]: I0124 01:04:39.366262 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 01:04:39 crc kubenswrapper[4676]: I0124 01:04:39.366372 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" gracePeriod=600 Jan 24 01:04:39 crc kubenswrapper[4676]: E0124 01:04:39.498037 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:04:40 crc kubenswrapper[4676]: I0124 01:04:40.444924 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" exitCode=0 Jan 24 01:04:40 crc kubenswrapper[4676]: I0124 01:04:40.445016 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9"} Jan 24 01:04:40 crc kubenswrapper[4676]: I0124 01:04:40.445248 4676 scope.go:117] "RemoveContainer" containerID="e910204e57fdf9a2ba78d40b8e9fd8506cf9f7184737c120d4af2680b61668e0" Jan 24 01:04:40 crc kubenswrapper[4676]: I0124 01:04:40.445943 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:04:40 crc kubenswrapper[4676]: E0124 01:04:40.446310 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.084013 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pgb89/must-gather-rzdhh"] Jan 24 01:04:41 crc kubenswrapper[4676]: E0124 01:04:41.084438 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerName="extract-utilities" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.084452 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerName="extract-utilities" Jan 24 01:04:41 crc kubenswrapper[4676]: E0124 01:04:41.084466 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerName="registry-server" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.084473 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerName="registry-server" Jan 24 01:04:41 crc kubenswrapper[4676]: E0124 01:04:41.084485 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerName="extract-content" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.084492 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerName="extract-content" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.084672 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="45ec6f03-36a9-4179-aee2-d3ba73569525" containerName="registry-server" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.085705 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.089344 4676 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pgb89"/"default-dockercfg-lvdrx" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.089511 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pgb89"/"kube-root-ca.crt" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.089642 4676 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pgb89"/"openshift-service-ca.crt" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.093843 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pgb89/must-gather-rzdhh"] Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.184565 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ce7f9d18-7eec-43f4-908a-d65d210492e7-must-gather-output\") pod \"must-gather-rzdhh\" (UID: \"ce7f9d18-7eec-43f4-908a-d65d210492e7\") " pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.184723 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/ce7f9d18-7eec-43f4-908a-d65d210492e7-kube-api-access-z2kcb\") pod \"must-gather-rzdhh\" (UID: \"ce7f9d18-7eec-43f4-908a-d65d210492e7\") " pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.286616 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/ce7f9d18-7eec-43f4-908a-d65d210492e7-kube-api-access-z2kcb\") pod \"must-gather-rzdhh\" (UID: \"ce7f9d18-7eec-43f4-908a-d65d210492e7\") " pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.286704 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ce7f9d18-7eec-43f4-908a-d65d210492e7-must-gather-output\") pod \"must-gather-rzdhh\" (UID: \"ce7f9d18-7eec-43f4-908a-d65d210492e7\") " pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.287142 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ce7f9d18-7eec-43f4-908a-d65d210492e7-must-gather-output\") pod \"must-gather-rzdhh\" (UID: \"ce7f9d18-7eec-43f4-908a-d65d210492e7\") " pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.310022 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/ce7f9d18-7eec-43f4-908a-d65d210492e7-kube-api-access-z2kcb\") pod \"must-gather-rzdhh\" (UID: \"ce7f9d18-7eec-43f4-908a-d65d210492e7\") " pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.452424 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:04:41 crc kubenswrapper[4676]: I0124 01:04:41.918679 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pgb89/must-gather-rzdhh"] Jan 24 01:04:42 crc kubenswrapper[4676]: I0124 01:04:42.482582 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pgb89/must-gather-rzdhh" event={"ID":"ce7f9d18-7eec-43f4-908a-d65d210492e7","Type":"ContainerStarted","Data":"ebdf94432be40adc9409ec2d7d078306d66cbfc22f6dc35636a6fecb28c07f9e"} Jan 24 01:04:42 crc kubenswrapper[4676]: I0124 01:04:42.482858 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pgb89/must-gather-rzdhh" event={"ID":"ce7f9d18-7eec-43f4-908a-d65d210492e7","Type":"ContainerStarted","Data":"d0b712dc713fcee1a9dcfc5340b4f493d3bd6dca5b50bbf686427d3022f3f332"} Jan 24 01:04:42 crc kubenswrapper[4676]: I0124 01:04:42.482872 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pgb89/must-gather-rzdhh" event={"ID":"ce7f9d18-7eec-43f4-908a-d65d210492e7","Type":"ContainerStarted","Data":"16c437b62ad21fcf66bdc37ab44a5e538e268edbe29d7c40000ad0f924ce60b1"} Jan 24 01:04:42 crc kubenswrapper[4676]: I0124 01:04:42.499737 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pgb89/must-gather-rzdhh" podStartSLOduration=1.499719873 podStartE2EDuration="1.499719873s" podCreationTimestamp="2026-01-24 01:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 01:04:42.4983185 +0000 UTC m=+3666.528289521" watchObservedRunningTime="2026-01-24 01:04:42.499719873 +0000 UTC m=+3666.529690874" Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.034027 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pgb89/crc-debug-5mpdp"] Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.036134 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.090156 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-host\") pod \"crc-debug-5mpdp\" (UID: \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\") " pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.090503 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tl8t\" (UniqueName: \"kubernetes.io/projected/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-kube-api-access-7tl8t\") pod \"crc-debug-5mpdp\" (UID: \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\") " pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.192704 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-host\") pod \"crc-debug-5mpdp\" (UID: \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\") " pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.192763 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tl8t\" (UniqueName: \"kubernetes.io/projected/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-kube-api-access-7tl8t\") pod \"crc-debug-5mpdp\" (UID: \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\") " pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.193059 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-host\") pod \"crc-debug-5mpdp\" (UID: \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\") " pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.228098 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tl8t\" (UniqueName: \"kubernetes.io/projected/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-kube-api-access-7tl8t\") pod \"crc-debug-5mpdp\" (UID: \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\") " pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.365409 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:04:46 crc kubenswrapper[4676]: W0124 01:04:46.421136 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2d41c9a_a05f_46c4_b548_fb6bfc0dee4a.slice/crio-f06001573ddb33783b3221d0ddb7f1268c8d975b7736cec8fc9e1555d436bf3d WatchSource:0}: Error finding container f06001573ddb33783b3221d0ddb7f1268c8d975b7736cec8fc9e1555d436bf3d: Status 404 returned error can't find the container with id f06001573ddb33783b3221d0ddb7f1268c8d975b7736cec8fc9e1555d436bf3d Jan 24 01:04:46 crc kubenswrapper[4676]: I0124 01:04:46.523289 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pgb89/crc-debug-5mpdp" event={"ID":"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a","Type":"ContainerStarted","Data":"f06001573ddb33783b3221d0ddb7f1268c8d975b7736cec8fc9e1555d436bf3d"} Jan 24 01:04:47 crc kubenswrapper[4676]: I0124 01:04:47.533055 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pgb89/crc-debug-5mpdp" event={"ID":"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a","Type":"ContainerStarted","Data":"6e1d712a5b60b5fed2dd683d40f7c0b75205c48776013f310795ca214bdf4a77"} Jan 24 01:04:47 crc kubenswrapper[4676]: I0124 01:04:47.548721 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pgb89/crc-debug-5mpdp" podStartSLOduration=1.5487052810000002 podStartE2EDuration="1.548705281s" podCreationTimestamp="2026-01-24 01:04:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-24 01:04:47.544820271 +0000 UTC m=+3671.574791272" watchObservedRunningTime="2026-01-24 01:04:47.548705281 +0000 UTC m=+3671.578676282" Jan 24 01:04:54 crc kubenswrapper[4676]: I0124 01:04:54.255555 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:04:54 crc kubenswrapper[4676]: E0124 01:04:54.256351 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:04:57 crc kubenswrapper[4676]: I0124 01:04:57.617669 4676 generic.go:334] "Generic (PLEG): container finished" podID="c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a" containerID="6e1d712a5b60b5fed2dd683d40f7c0b75205c48776013f310795ca214bdf4a77" exitCode=0 Jan 24 01:04:57 crc kubenswrapper[4676]: I0124 01:04:57.617753 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pgb89/crc-debug-5mpdp" event={"ID":"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a","Type":"ContainerDied","Data":"6e1d712a5b60b5fed2dd683d40f7c0b75205c48776013f310795ca214bdf4a77"} Jan 24 01:04:58 crc kubenswrapper[4676]: I0124 01:04:58.762718 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:04:58 crc kubenswrapper[4676]: I0124 01:04:58.814708 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pgb89/crc-debug-5mpdp"] Jan 24 01:04:58 crc kubenswrapper[4676]: I0124 01:04:58.821627 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pgb89/crc-debug-5mpdp"] Jan 24 01:04:58 crc kubenswrapper[4676]: I0124 01:04:58.824492 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tl8t\" (UniqueName: \"kubernetes.io/projected/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-kube-api-access-7tl8t\") pod \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\" (UID: \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\") " Jan 24 01:04:58 crc kubenswrapper[4676]: I0124 01:04:58.824603 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-host\") pod \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\" (UID: \"c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a\") " Jan 24 01:04:58 crc kubenswrapper[4676]: I0124 01:04:58.825021 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-host" (OuterVolumeSpecName: "host") pod "c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a" (UID: "c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 01:04:58 crc kubenswrapper[4676]: I0124 01:04:58.834608 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-kube-api-access-7tl8t" (OuterVolumeSpecName: "kube-api-access-7tl8t") pod "c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a" (UID: "c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a"). InnerVolumeSpecName "kube-api-access-7tl8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:04:58 crc kubenswrapper[4676]: I0124 01:04:58.926737 4676 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-host\") on node \"crc\" DevicePath \"\"" Jan 24 01:04:58 crc kubenswrapper[4676]: I0124 01:04:58.927029 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tl8t\" (UniqueName: \"kubernetes.io/projected/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a-kube-api-access-7tl8t\") on node \"crc\" DevicePath \"\"" Jan 24 01:04:59 crc kubenswrapper[4676]: I0124 01:04:59.635820 4676 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f06001573ddb33783b3221d0ddb7f1268c8d975b7736cec8fc9e1555d436bf3d" Jan 24 01:04:59 crc kubenswrapper[4676]: I0124 01:04:59.636178 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/crc-debug-5mpdp" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.102068 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pgb89/crc-debug-vxdn7"] Jan 24 01:05:00 crc kubenswrapper[4676]: E0124 01:05:00.103462 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a" containerName="container-00" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.103558 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a" containerName="container-00" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.103803 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a" containerName="container-00" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.104475 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.150157 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-host\") pod \"crc-debug-vxdn7\" (UID: \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\") " pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.150236 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqnvr\" (UniqueName: \"kubernetes.io/projected/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-kube-api-access-xqnvr\") pod \"crc-debug-vxdn7\" (UID: \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\") " pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.252813 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-host\") pod \"crc-debug-vxdn7\" (UID: \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\") " pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.252867 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqnvr\" (UniqueName: \"kubernetes.io/projected/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-kube-api-access-xqnvr\") pod \"crc-debug-vxdn7\" (UID: \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\") " pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.252974 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-host\") pod \"crc-debug-vxdn7\" (UID: \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\") " pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.266677 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a" path="/var/lib/kubelet/pods/c2d41c9a-a05f-46c4-b548-fb6bfc0dee4a/volumes" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.282844 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqnvr\" (UniqueName: \"kubernetes.io/projected/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-kube-api-access-xqnvr\") pod \"crc-debug-vxdn7\" (UID: \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\") " pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.430552 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:00 crc kubenswrapper[4676]: W0124 01:05:00.481624 4676 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd98ed69b_2cc8_4c59_a562_d1cafcfcfcf8.slice/crio-7332cafa0ac25e2d33eeed9dc137e087b955765c89b55c9d3fff5e9dc7f907c0 WatchSource:0}: Error finding container 7332cafa0ac25e2d33eeed9dc137e087b955765c89b55c9d3fff5e9dc7f907c0: Status 404 returned error can't find the container with id 7332cafa0ac25e2d33eeed9dc137e087b955765c89b55c9d3fff5e9dc7f907c0 Jan 24 01:05:00 crc kubenswrapper[4676]: I0124 01:05:00.645058 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pgb89/crc-debug-vxdn7" event={"ID":"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8","Type":"ContainerStarted","Data":"7332cafa0ac25e2d33eeed9dc137e087b955765c89b55c9d3fff5e9dc7f907c0"} Jan 24 01:05:01 crc kubenswrapper[4676]: I0124 01:05:01.655213 4676 generic.go:334] "Generic (PLEG): container finished" podID="d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8" containerID="a96dc5b6f6afa933b2ccb20288d52f9f02e4f6788c82a166648f646efd76a1ef" exitCode=1 Jan 24 01:05:01 crc kubenswrapper[4676]: I0124 01:05:01.655430 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pgb89/crc-debug-vxdn7" event={"ID":"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8","Type":"ContainerDied","Data":"a96dc5b6f6afa933b2ccb20288d52f9f02e4f6788c82a166648f646efd76a1ef"} Jan 24 01:05:01 crc kubenswrapper[4676]: I0124 01:05:01.698959 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pgb89/crc-debug-vxdn7"] Jan 24 01:05:01 crc kubenswrapper[4676]: I0124 01:05:01.706293 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pgb89/crc-debug-vxdn7"] Jan 24 01:05:02 crc kubenswrapper[4676]: I0124 01:05:02.752615 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:02 crc kubenswrapper[4676]: I0124 01:05:02.796568 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-host\") pod \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\" (UID: \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\") " Jan 24 01:05:02 crc kubenswrapper[4676]: I0124 01:05:02.796846 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqnvr\" (UniqueName: \"kubernetes.io/projected/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-kube-api-access-xqnvr\") pod \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\" (UID: \"d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8\") " Jan 24 01:05:02 crc kubenswrapper[4676]: I0124 01:05:02.796695 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-host" (OuterVolumeSpecName: "host") pod "d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8" (UID: "d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 24 01:05:02 crc kubenswrapper[4676]: I0124 01:05:02.814078 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-kube-api-access-xqnvr" (OuterVolumeSpecName: "kube-api-access-xqnvr") pod "d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8" (UID: "d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8"). InnerVolumeSpecName "kube-api-access-xqnvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:05:02 crc kubenswrapper[4676]: I0124 01:05:02.898929 4676 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-host\") on node \"crc\" DevicePath \"\"" Jan 24 01:05:02 crc kubenswrapper[4676]: I0124 01:05:02.898956 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqnvr\" (UniqueName: \"kubernetes.io/projected/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8-kube-api-access-xqnvr\") on node \"crc\" DevicePath \"\"" Jan 24 01:05:03 crc kubenswrapper[4676]: I0124 01:05:03.673087 4676 scope.go:117] "RemoveContainer" containerID="a96dc5b6f6afa933b2ccb20288d52f9f02e4f6788c82a166648f646efd76a1ef" Jan 24 01:05:03 crc kubenswrapper[4676]: I0124 01:05:03.673166 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/crc-debug-vxdn7" Jan 24 01:05:04 crc kubenswrapper[4676]: I0124 01:05:04.279038 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8" path="/var/lib/kubelet/pods/d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8/volumes" Jan 24 01:05:09 crc kubenswrapper[4676]: I0124 01:05:09.256228 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:05:09 crc kubenswrapper[4676]: E0124 01:05:09.257050 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:05:21 crc kubenswrapper[4676]: I0124 01:05:21.255698 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:05:21 crc kubenswrapper[4676]: E0124 01:05:21.256991 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:05:36 crc kubenswrapper[4676]: I0124 01:05:36.277225 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:05:36 crc kubenswrapper[4676]: E0124 01:05:36.279114 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:05:44 crc kubenswrapper[4676]: I0124 01:05:44.418541 4676 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-f47776b4c-v4xb2" podUID="b976b9e2-b80e-4626-919d-3bb84f0151e8" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Jan 24 01:05:47 crc kubenswrapper[4676]: I0124 01:05:47.255444 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:05:47 crc kubenswrapper[4676]: E0124 01:05:47.256075 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:06:00 crc kubenswrapper[4676]: I0124 01:06:00.256593 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:06:00 crc kubenswrapper[4676]: E0124 01:06:00.257366 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:06:00 crc kubenswrapper[4676]: I0124 01:06:00.565077 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54b7b855f4-s49zw_c015646b-9361-4b5c-b465-a66d7fc5cc53/barbican-api/0.log" Jan 24 01:06:00 crc kubenswrapper[4676]: I0124 01:06:00.737823 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-54b7b855f4-s49zw_c015646b-9361-4b5c-b465-a66d7fc5cc53/barbican-api-log/0.log" Jan 24 01:06:00 crc kubenswrapper[4676]: I0124 01:06:00.839691 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79ccd69c74-nj8k6_a05052ce-062e-423c-80cf-78349e42718f/barbican-keystone-listener/0.log" Jan 24 01:06:00 crc kubenswrapper[4676]: I0124 01:06:00.890137 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-79ccd69c74-nj8k6_a05052ce-062e-423c-80cf-78349e42718f/barbican-keystone-listener-log/0.log" Jan 24 01:06:01 crc kubenswrapper[4676]: I0124 01:06:01.007825 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68bd7fb46c-sflbz_009f35e0-3a98-453b-b92e-8db9e5c92798/barbican-worker/0.log" Jan 24 01:06:01 crc kubenswrapper[4676]: I0124 01:06:01.075288 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-68bd7fb46c-sflbz_009f35e0-3a98-453b-b92e-8db9e5c92798/barbican-worker-log/0.log" Jan 24 01:06:01 crc kubenswrapper[4676]: I0124 01:06:01.194425 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-9z5hp_1fd4e1f4-0772-493a-b929-6e93470f9abf/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:01 crc kubenswrapper[4676]: I0124 01:06:01.284841 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58bbd77a-9518-4037-b96b-a1490082fb04/ceilometer-central-agent/0.log" Jan 24 01:06:01 crc kubenswrapper[4676]: I0124 01:06:01.869838 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58bbd77a-9518-4037-b96b-a1490082fb04/sg-core/0.log" Jan 24 01:06:01 crc kubenswrapper[4676]: I0124 01:06:01.894012 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58bbd77a-9518-4037-b96b-a1490082fb04/ceilometer-notification-agent/0.log" Jan 24 01:06:01 crc kubenswrapper[4676]: I0124 01:06:01.901158 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_58bbd77a-9518-4037-b96b-a1490082fb04/proxy-httpd/0.log" Jan 24 01:06:02 crc kubenswrapper[4676]: I0124 01:06:02.099992 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d38cef52-387b-4633-be75-6dc455ad53c4/cinder-api/0.log" Jan 24 01:06:02 crc kubenswrapper[4676]: I0124 01:06:02.130246 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d38cef52-387b-4633-be75-6dc455ad53c4/cinder-api-log/0.log" Jan 24 01:06:02 crc kubenswrapper[4676]: I0124 01:06:02.272725 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978/cinder-scheduler/0.log" Jan 24 01:06:02 crc kubenswrapper[4676]: I0124 01:06:02.311517 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_629a8ba3-3e4e-4fc2-86a4-bb07d2fb4978/probe/0.log" Jan 24 01:06:02 crc kubenswrapper[4676]: I0124 01:06:02.418733 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-wwcsg_7f552002-93ef-485f-9227-a94733534466/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:02 crc kubenswrapper[4676]: I0124 01:06:02.576877 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-fb7f7_f6bc5ee4-f730-4e1e-9684-b643daed2519/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:03 crc kubenswrapper[4676]: I0124 01:06:03.301482 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-drclt_cdf394f7-5d67-4a0f-9644-82fe83a72e2d/init/0.log" Jan 24 01:06:03 crc kubenswrapper[4676]: I0124 01:06:03.446251 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-drclt_cdf394f7-5d67-4a0f-9644-82fe83a72e2d/init/0.log" Jan 24 01:06:03 crc kubenswrapper[4676]: I0124 01:06:03.480560 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6b865b64bc-drclt_cdf394f7-5d67-4a0f-9644-82fe83a72e2d/dnsmasq-dns/0.log" Jan 24 01:06:03 crc kubenswrapper[4676]: I0124 01:06:03.547695 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-skh8k_7e20d58a-5f01-4c23-9ab0-650d3ae76844/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:03 crc kubenswrapper[4676]: I0124 01:06:03.722895 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a6cef518-8385-4315-83d5-f46a6144d5a0/glance-log/0.log" Jan 24 01:06:03 crc kubenswrapper[4676]: I0124 01:06:03.778455 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a6cef518-8385-4315-83d5-f46a6144d5a0/glance-httpd/0.log" Jan 24 01:06:03 crc kubenswrapper[4676]: I0124 01:06:03.889945 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_042ec835-b8c1-43be-a19d-d70f76128e26/glance-httpd/0.log" Jan 24 01:06:03 crc kubenswrapper[4676]: I0124 01:06:03.925991 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_042ec835-b8c1-43be-a19d-d70f76128e26/glance-log/0.log" Jan 24 01:06:04 crc kubenswrapper[4676]: I0124 01:06:04.041703 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f876ddf46-fs7qv_ac7dce6b-3bd9-4ad9-9485-83d9384b8bad/horizon/1.log" Jan 24 01:06:04 crc kubenswrapper[4676]: I0124 01:06:04.135119 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f876ddf46-fs7qv_ac7dce6b-3bd9-4ad9-9485-83d9384b8bad/horizon/0.log" Jan 24 01:06:04 crc kubenswrapper[4676]: I0124 01:06:04.376688 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-t4nrs_e9d2a92f-e22e-44c5-86c5-8b38824e3d4c/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:04 crc kubenswrapper[4676]: I0124 01:06:04.423012 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-f876ddf46-fs7qv_ac7dce6b-3bd9-4ad9-9485-83d9384b8bad/horizon-log/0.log" Jan 24 01:06:04 crc kubenswrapper[4676]: I0124 01:06:04.592075 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-sz24c_15fbef8d-e606-4b72-a994-e71d03e8fec8/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:04 crc kubenswrapper[4676]: I0124 01:06:04.704999 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-849b597d57-kw79c_29bf9e4b-4fb3-41f4-9280-f7ea2e61a844/keystone-api/0.log" Jan 24 01:06:04 crc kubenswrapper[4676]: I0124 01:06:04.764370 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29486941-6wtvm_483961de-208d-4593-a6c8-ecee687b7c06/keystone-cron/0.log" Jan 24 01:06:05 crc kubenswrapper[4676]: I0124 01:06:05.036612 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_25ee749c-0b84-4abd-9fe0-a6f23c0c912d/kube-state-metrics/0.log" Jan 24 01:06:05 crc kubenswrapper[4676]: I0124 01:06:05.143582 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-ddk99_2cafe497-da96-4b39-bec2-1ec54f859303/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:05 crc kubenswrapper[4676]: I0124 01:06:05.235492 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_545f6045-cf2f-4d4b-91d8-227148ddd71a/memcached/0.log" Jan 24 01:06:05 crc kubenswrapper[4676]: I0124 01:06:05.386676 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd579cfd9-q4npp_6dda8455-6777-4efc-abf3-df547cf58339/neutron-api/0.log" Jan 24 01:06:05 crc kubenswrapper[4676]: I0124 01:06:05.569328 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bd579cfd9-q4npp_6dda8455-6777-4efc-abf3-df547cf58339/neutron-httpd/0.log" Jan 24 01:06:05 crc kubenswrapper[4676]: I0124 01:06:05.691128 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-92xr7_cba640eb-c65f-46be-af5d-5126418c361a/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:05 crc kubenswrapper[4676]: I0124 01:06:05.987709 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ff542eb3-0142-4d46-a6e6-2e89c73f5824/nova-api-api/0.log" Jan 24 01:06:06 crc kubenswrapper[4676]: I0124 01:06:06.100969 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_ff542eb3-0142-4d46-a6e6-2e89c73f5824/nova-api-log/0.log" Jan 24 01:06:06 crc kubenswrapper[4676]: I0124 01:06:06.124136 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_400ba963-913a-401c-8f2e-21005977e0c2/nova-cell0-conductor-conductor/0.log" Jan 24 01:06:06 crc kubenswrapper[4676]: I0124 01:06:06.359458 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_4420d197-3908-406b-9661-67d64ccd7768/nova-cell1-conductor-conductor/0.log" Jan 24 01:06:06 crc kubenswrapper[4676]: I0124 01:06:06.460678 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_b1499ed8-1041-4980-b9da-cb957cbf215c/nova-cell1-novncproxy-novncproxy/0.log" Jan 24 01:06:06 crc kubenswrapper[4676]: I0124 01:06:06.924735 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-5cl2g_75cae94d-b818-4f92-b42d-fd8cec63a657/nova-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.081636 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1085d48-78f1-4437-a518-239ba90b7c0b/nova-metadata-log/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.391480 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_19365292-50d8-4e94-952f-2df7ee20f0ba/mysql-bootstrap/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.465305 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_8ff6f87b-a571-43da-9fbc-9203f0001771/nova-scheduler-scheduler/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.614735 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_19365292-50d8-4e94-952f-2df7ee20f0ba/mysql-bootstrap/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.643341 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f1085d48-78f1-4437-a518-239ba90b7c0b/nova-metadata-metadata/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.661304 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_19365292-50d8-4e94-952f-2df7ee20f0ba/galera/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.704751 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2bbaae64-ac2d-43c6-8984-5483f2eb4211/mysql-bootstrap/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.863885 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2bbaae64-ac2d-43c6-8984-5483f2eb4211/mysql-bootstrap/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.901559 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2bbaae64-ac2d-43c6-8984-5483f2eb4211/galera/0.log" Jan 24 01:06:07 crc kubenswrapper[4676]: I0124 01:06:07.901600 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f3fa6e59-785d-4d40-8d73-170552068e43/openstackclient/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.109313 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jsd4n_7adbcf83-efbd-4e8d-97e5-f8768463284a/ovn-controller/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.151362 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-mq6ns_21550436-cd71-46d8-838e-c51c19ddf8ff/openstack-network-exporter/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.239040 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6sl9q_427fbd2d-16ef-44a6-a71d-8172f56b863d/ovsdb-server-init/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.440715 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6sl9q_427fbd2d-16ef-44a6-a71d-8172f56b863d/ovsdb-server-init/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.445175 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6sl9q_427fbd2d-16ef-44a6-a71d-8172f56b863d/ovsdb-server/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.505126 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-527bv_55444bfa-a024-4606-aa57-6456c6688e52/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.514033 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-6sl9q_427fbd2d-16ef-44a6-a71d-8172f56b863d/ovs-vswitchd/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.649091 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1486ea92-d267-49cd-8516-d474ef25c2df/openstack-network-exporter/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.709719 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_1486ea92-d267-49cd-8516-d474ef25c2df/ovn-northd/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.824793 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3bff778a-b10f-4ba9-a12f-f4086608fd30/openstack-network-exporter/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.875749 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_3bff778a-b10f-4ba9-a12f-f4086608fd30/ovsdbserver-nb/0.log" Jan 24 01:06:08 crc kubenswrapper[4676]: I0124 01:06:08.944717 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f3648c65-9fcb-4a9e-b4cb-d8437dc00141/openstack-network-exporter/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.134321 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f3648c65-9fcb-4a9e-b4cb-d8437dc00141/ovsdbserver-sb/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.136583 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-765f6cdf58-5q9v9_21e63383-223a-4247-8589-03ab5a33f980/placement-api/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.198061 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-765f6cdf58-5q9v9_21e63383-223a-4247-8589-03ab5a33f980/placement-log/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.295019 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a2df1d42-fa93-4771-ba77-1c27f820b298/setup-container/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.435868 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a2df1d42-fa93-4771-ba77-1c27f820b298/rabbitmq/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.470219 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a2df1d42-fa93-4771-ba77-1c27f820b298/setup-container/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.492696 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c162e478-58e3-4a83-97cb-29887613c1aa/setup-container/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.681845 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c162e478-58e3-4a83-97cb-29887613c1aa/setup-container/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.721786 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c162e478-58e3-4a83-97cb-29887613c1aa/rabbitmq/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.726768 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-448ht_447c1e1f-d798-4bcc-a8ef-91d4ad5426a5/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.918030 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qx6r8_19b712ec-28a7-419f-9f09-7d0b0ecbf747/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:09 crc kubenswrapper[4676]: I0124 01:06:09.970006 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-fwvvd_15cfbdc6-1f3b-49a5-8822-c4af1e686731/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.086467 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-mhrvn_2fc6ead9-bdd0-49ef-9da8-96ccf67f6ec1/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.173271 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8b9rz_d1ed73dc-4392-4d20-a592-4a8c5ba9c104/ssh-known-hosts-edpm-deployment/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.406718 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f47776b4c-v4xb2_b976b9e2-b80e-4626-919d-3bb84f0151e8/proxy-httpd/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.465803 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f47776b4c-v4xb2_b976b9e2-b80e-4626-919d-3bb84f0151e8/proxy-server/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.553095 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-5fmzb_4ff61f48-e451-47e8-adcc-0870b29d28a9/swift-ring-rebalance/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.671757 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/account-reaper/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.693775 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/account-auditor/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.738665 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/account-replicator/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.785595 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/account-server/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.816353 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/container-auditor/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.916597 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/container-replicator/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.926936 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/container-server/0.log" Jan 24 01:06:10 crc kubenswrapper[4676]: I0124 01:06:10.963707 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/container-updater/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.007545 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-auditor/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.017080 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-expirer/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.114800 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-server/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.147834 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-replicator/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.163843 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/object-updater/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.226815 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/rsync/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.228166 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_4620e725-9218-461b-a56d-104bcb7f1df4/swift-recon-cron/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.419784 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_3e2adf44-9053-4dcd-9d47-27910710dbc8/tempest-tests-tempest-tests-runner/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.458017 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-j8gtv_7ef31551-e4ed-48d0-a4d6-f9c2fb515966/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.559412 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_e1574f42-e89a-40d4-b6da-2d4ef0824916/test-operator-logs-container/0.log" Jan 24 01:06:11 crc kubenswrapper[4676]: I0124 01:06:11.686047 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-hx5pd_83640815-cc06-4abe-a06f-20a1f8798609/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Jan 24 01:06:13 crc kubenswrapper[4676]: I0124 01:06:13.256011 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:06:13 crc kubenswrapper[4676]: E0124 01:06:13.256647 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.267792 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7cbcm"] Jan 24 01:06:20 crc kubenswrapper[4676]: E0124 01:06:20.268750 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8" containerName="container-00" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.268767 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8" containerName="container-00" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.268993 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="d98ed69b-2cc8-4c59-a562-d1cafcfcfcf8" containerName="container-00" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.270387 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.285611 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7cbcm"] Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.375387 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-catalog-content\") pod \"community-operators-7cbcm\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.375556 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kxvg\" (UniqueName: \"kubernetes.io/projected/218ff773-95bb-4b63-a4d5-b7fb752b5870-kube-api-access-7kxvg\") pod \"community-operators-7cbcm\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.375590 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-utilities\") pod \"community-operators-7cbcm\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.477305 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kxvg\" (UniqueName: \"kubernetes.io/projected/218ff773-95bb-4b63-a4d5-b7fb752b5870-kube-api-access-7kxvg\") pod \"community-operators-7cbcm\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.477356 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-utilities\") pod \"community-operators-7cbcm\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.477450 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-catalog-content\") pod \"community-operators-7cbcm\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.477842 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-catalog-content\") pod \"community-operators-7cbcm\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.478305 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-utilities\") pod \"community-operators-7cbcm\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.497295 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kxvg\" (UniqueName: \"kubernetes.io/projected/218ff773-95bb-4b63-a4d5-b7fb752b5870-kube-api-access-7kxvg\") pod \"community-operators-7cbcm\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:20 crc kubenswrapper[4676]: I0124 01:06:20.627114 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:21 crc kubenswrapper[4676]: I0124 01:06:21.183334 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7cbcm"] Jan 24 01:06:21 crc kubenswrapper[4676]: I0124 01:06:21.539568 4676 generic.go:334] "Generic (PLEG): container finished" podID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerID="d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06" exitCode=0 Jan 24 01:06:21 crc kubenswrapper[4676]: I0124 01:06:21.539670 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cbcm" event={"ID":"218ff773-95bb-4b63-a4d5-b7fb752b5870","Type":"ContainerDied","Data":"d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06"} Jan 24 01:06:21 crc kubenswrapper[4676]: I0124 01:06:21.539808 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cbcm" event={"ID":"218ff773-95bb-4b63-a4d5-b7fb752b5870","Type":"ContainerStarted","Data":"76f1d671e1f62de1005f97f3b82a4e50a1c9871c76ad758d00aea03a6ba36366"} Jan 24 01:06:23 crc kubenswrapper[4676]: I0124 01:06:23.556761 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cbcm" event={"ID":"218ff773-95bb-4b63-a4d5-b7fb752b5870","Type":"ContainerStarted","Data":"c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50"} Jan 24 01:06:24 crc kubenswrapper[4676]: I0124 01:06:24.566705 4676 generic.go:334] "Generic (PLEG): container finished" podID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerID="c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50" exitCode=0 Jan 24 01:06:24 crc kubenswrapper[4676]: I0124 01:06:24.566800 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cbcm" event={"ID":"218ff773-95bb-4b63-a4d5-b7fb752b5870","Type":"ContainerDied","Data":"c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50"} Jan 24 01:06:24 crc kubenswrapper[4676]: I0124 01:06:24.567199 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cbcm" event={"ID":"218ff773-95bb-4b63-a4d5-b7fb752b5870","Type":"ContainerStarted","Data":"4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330"} Jan 24 01:06:24 crc kubenswrapper[4676]: I0124 01:06:24.584094 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7cbcm" podStartSLOduration=2.064078141 podStartE2EDuration="4.584077772s" podCreationTimestamp="2026-01-24 01:06:20 +0000 UTC" firstStartedPulling="2026-01-24 01:06:21.540777145 +0000 UTC m=+3765.570748146" lastFinishedPulling="2026-01-24 01:06:24.060776776 +0000 UTC m=+3768.090747777" observedRunningTime="2026-01-24 01:06:24.579784531 +0000 UTC m=+3768.609755532" watchObservedRunningTime="2026-01-24 01:06:24.584077772 +0000 UTC m=+3768.614048773" Jan 24 01:06:25 crc kubenswrapper[4676]: I0124 01:06:25.255737 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:06:25 crc kubenswrapper[4676]: E0124 01:06:25.256104 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:06:30 crc kubenswrapper[4676]: I0124 01:06:30.640525 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:30 crc kubenswrapper[4676]: I0124 01:06:30.642265 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:31 crc kubenswrapper[4676]: I0124 01:06:31.063660 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:31 crc kubenswrapper[4676]: I0124 01:06:31.699470 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:31 crc kubenswrapper[4676]: I0124 01:06:31.747140 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7cbcm"] Jan 24 01:06:33 crc kubenswrapper[4676]: I0124 01:06:33.667707 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7cbcm" podUID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerName="registry-server" containerID="cri-o://4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330" gracePeriod=2 Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.633948 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.677745 4676 generic.go:334] "Generic (PLEG): container finished" podID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerID="4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330" exitCode=0 Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.677788 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cbcm" event={"ID":"218ff773-95bb-4b63-a4d5-b7fb752b5870","Type":"ContainerDied","Data":"4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330"} Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.677818 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cbcm" event={"ID":"218ff773-95bb-4b63-a4d5-b7fb752b5870","Type":"ContainerDied","Data":"76f1d671e1f62de1005f97f3b82a4e50a1c9871c76ad758d00aea03a6ba36366"} Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.677836 4676 scope.go:117] "RemoveContainer" containerID="4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.677986 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cbcm" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.699286 4676 scope.go:117] "RemoveContainer" containerID="c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.716919 4676 scope.go:117] "RemoveContainer" containerID="d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.752686 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-catalog-content\") pod \"218ff773-95bb-4b63-a4d5-b7fb752b5870\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.752858 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-utilities\") pod \"218ff773-95bb-4b63-a4d5-b7fb752b5870\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.752902 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kxvg\" (UniqueName: \"kubernetes.io/projected/218ff773-95bb-4b63-a4d5-b7fb752b5870-kube-api-access-7kxvg\") pod \"218ff773-95bb-4b63-a4d5-b7fb752b5870\" (UID: \"218ff773-95bb-4b63-a4d5-b7fb752b5870\") " Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.754751 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-utilities" (OuterVolumeSpecName: "utilities") pod "218ff773-95bb-4b63-a4d5-b7fb752b5870" (UID: "218ff773-95bb-4b63-a4d5-b7fb752b5870"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.760672 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/218ff773-95bb-4b63-a4d5-b7fb752b5870-kube-api-access-7kxvg" (OuterVolumeSpecName: "kube-api-access-7kxvg") pod "218ff773-95bb-4b63-a4d5-b7fb752b5870" (UID: "218ff773-95bb-4b63-a4d5-b7fb752b5870"). InnerVolumeSpecName "kube-api-access-7kxvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.771706 4676 scope.go:117] "RemoveContainer" containerID="4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330" Jan 24 01:06:34 crc kubenswrapper[4676]: E0124 01:06:34.783840 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330\": container with ID starting with 4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330 not found: ID does not exist" containerID="4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.783875 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330"} err="failed to get container status \"4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330\": rpc error: code = NotFound desc = could not find container \"4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330\": container with ID starting with 4623a790b30d939fdb7ee79f3f467c46f1f8092f29d016a602d7699d5ad76330 not found: ID does not exist" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.783899 4676 scope.go:117] "RemoveContainer" containerID="c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50" Jan 24 01:06:34 crc kubenswrapper[4676]: E0124 01:06:34.784341 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50\": container with ID starting with c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50 not found: ID does not exist" containerID="c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.784422 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50"} err="failed to get container status \"c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50\": rpc error: code = NotFound desc = could not find container \"c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50\": container with ID starting with c086e3b7dcaf62c1a5d2423eca56792f5f7e8612c92104aeb1a3622b5b0feb50 not found: ID does not exist" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.784453 4676 scope.go:117] "RemoveContainer" containerID="d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06" Jan 24 01:06:34 crc kubenswrapper[4676]: E0124 01:06:34.785285 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06\": container with ID starting with d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06 not found: ID does not exist" containerID="d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.785313 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06"} err="failed to get container status \"d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06\": rpc error: code = NotFound desc = could not find container \"d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06\": container with ID starting with d2ceec6e32e43b9f2fd6d261e6b3cdd5b94e32ac761c00f5b4b557088654fd06 not found: ID does not exist" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.822219 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "218ff773-95bb-4b63-a4d5-b7fb752b5870" (UID: "218ff773-95bb-4b63-a4d5-b7fb752b5870"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.855320 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.855349 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kxvg\" (UniqueName: \"kubernetes.io/projected/218ff773-95bb-4b63-a4d5-b7fb752b5870-kube-api-access-7kxvg\") on node \"crc\" DevicePath \"\"" Jan 24 01:06:34 crc kubenswrapper[4676]: I0124 01:06:34.855360 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218ff773-95bb-4b63-a4d5-b7fb752b5870-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 01:06:35 crc kubenswrapper[4676]: I0124 01:06:35.025012 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7cbcm"] Jan 24 01:06:35 crc kubenswrapper[4676]: I0124 01:06:35.035136 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7cbcm"] Jan 24 01:06:36 crc kubenswrapper[4676]: I0124 01:06:36.275170 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:06:36 crc kubenswrapper[4676]: I0124 01:06:36.275365 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="218ff773-95bb-4b63-a4d5-b7fb752b5870" path="/var/lib/kubelet/pods/218ff773-95bb-4b63-a4d5-b7fb752b5870/volumes" Jan 24 01:06:36 crc kubenswrapper[4676]: E0124 01:06:36.275683 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:06:39 crc kubenswrapper[4676]: I0124 01:06:39.174916 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/util/0.log" Jan 24 01:06:39 crc kubenswrapper[4676]: I0124 01:06:39.291932 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/util/0.log" Jan 24 01:06:39 crc kubenswrapper[4676]: I0124 01:06:39.362602 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/pull/0.log" Jan 24 01:06:39 crc kubenswrapper[4676]: I0124 01:06:39.379300 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/pull/0.log" Jan 24 01:06:39 crc kubenswrapper[4676]: I0124 01:06:39.569345 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/pull/0.log" Jan 24 01:06:39 crc kubenswrapper[4676]: I0124 01:06:39.617559 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/util/0.log" Jan 24 01:06:39 crc kubenswrapper[4676]: I0124 01:06:39.632793 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ba3e01239114c6476dd7da80952015a1e39aad88ba09ffb7f3cf48a9e6vf789_9d669669-26ba-4775-b7cc-e97cc7dbe326/extract/0.log" Jan 24 01:06:39 crc kubenswrapper[4676]: I0124 01:06:39.842297 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7f86f8796f-fgplq_02123851-7d2f-477b-9c60-5a9922a0bc97/manager/0.log" Jan 24 01:06:39 crc kubenswrapper[4676]: I0124 01:06:39.892853 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-69cf5d4557-c8c6m_0cd05b9f-6699-46e3-ae36-9f21352e6c8e/manager/0.log" Jan 24 01:06:40 crc kubenswrapper[4676]: I0124 01:06:40.034407 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-js6db_a9f1e2a4-c9fa-4136-aa76-059dc2ed9c85/manager/0.log" Jan 24 01:06:40 crc kubenswrapper[4676]: I0124 01:06:40.187873 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-btxnv_29e4b64d-19bd-419b-9e29-7a41e6f12ae0/manager/0.log" Jan 24 01:06:40 crc kubenswrapper[4676]: I0124 01:06:40.319454 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-p2nr8_5e9cf1cb-c413-45ad-8a51-bf35407fcdfe/manager/0.log" Jan 24 01:06:40 crc kubenswrapper[4676]: I0124 01:06:40.446699 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-rbzsj_6b8541f9-a37a-41d6-8006-3d0335c3abb5/manager/0.log" Jan 24 01:06:40 crc kubenswrapper[4676]: I0124 01:06:40.745019 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-58749ffdfb-jxx26_921e121c-5261-4fe7-8171-6b634babedf4/manager/0.log" Jan 24 01:06:41 crc kubenswrapper[4676]: I0124 01:06:41.000844 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-598f7747c9-v25h4_555ebb8f-1bc3-4b8d-9f37-cad92b48477c/manager/0.log" Jan 24 01:06:41 crc kubenswrapper[4676]: I0124 01:06:41.198782 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-2gzqg_9a3f9a14-1138-425d-8a56-454b282d7d9f/manager/0.log" Jan 24 01:06:41 crc kubenswrapper[4676]: I0124 01:06:41.277465 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-g4d8z_dfc79179-d245-4360-be6e-8b43441e23ed/manager/0.log" Jan 24 01:06:41 crc kubenswrapper[4676]: I0124 01:06:41.482691 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6b9fb5fdcb-rpn8n_4ce661f6-26e2-4da2-a759-e493a60587b2/manager/0.log" Jan 24 01:06:41 crc kubenswrapper[4676]: I0124 01:06:41.609343 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78d58447c5-wqbcz_dd6346d8-9cf1-4364-b480-f4c2d872472f/manager/0.log" Jan 24 01:06:41 crc kubenswrapper[4676]: I0124 01:06:41.725618 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b8bc8d87d-4xp45_df9ab5f0-f577-4303-8045-f960c67a6936/manager/0.log" Jan 24 01:06:41 crc kubenswrapper[4676]: I0124 01:06:41.851322 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7bd9774b6-k8lw7_ced74bcb-8345-40c5-b2d4-3d369f30b835/manager/0.log" Jan 24 01:06:42 crc kubenswrapper[4676]: I0124 01:06:42.002245 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b8549w5ms_196f45b9-e656-4760-b058-e0b5c08a50d9/manager/0.log" Jan 24 01:06:42 crc kubenswrapper[4676]: I0124 01:06:42.239154 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-69647cdbc5-96fbr_07dc00eb-bfcb-4d0d-bd6a-9e4b52e3e7f6/operator/0.log" Jan 24 01:06:42 crc kubenswrapper[4676]: I0124 01:06:42.449138 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-kw9l8_25146f99-405a-4473-bf27-69a7195a3338/registry-server/0.log" Jan 24 01:06:42 crc kubenswrapper[4676]: I0124 01:06:42.838352 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5d646b7d76-tdqwh_d85fa79d-818f-4079-aac4-f3fa51a90e9a/manager/0.log" Jan 24 01:06:42 crc kubenswrapper[4676]: I0124 01:06:42.918281 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-55db956ddc-49nrq_53358678-d763-4b02-a157-86a57ebd0305/manager/0.log" Jan 24 01:06:43 crc kubenswrapper[4676]: I0124 01:06:43.252143 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5d5f8c4f48-md6zs_9d79c791-c851-4c4a-aa2d-d175b668b0f5/manager/0.log" Jan 24 01:06:43 crc kubenswrapper[4676]: I0124 01:06:43.257759 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-h44qw_fce4c8d0-b903-4873-8c89-2f4b9dd9c05d/operator/0.log" Jan 24 01:06:43 crc kubenswrapper[4676]: I0124 01:06:43.342197 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5df95d5965-h8wx9_b0c8972b-31d7-40c1-bc65-1478718d41a5/manager/0.log" Jan 24 01:06:43 crc kubenswrapper[4676]: I0124 01:06:43.507136 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-4c5zl_060e1c8d-dfa6-428f-bffe-d89ac3dab8c3/manager/0.log" Jan 24 01:06:43 crc kubenswrapper[4676]: I0124 01:06:43.593513 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-gqg82_6245b73e-9fba-4ad7-bbbc-31db48c03825/manager/0.log" Jan 24 01:06:43 crc kubenswrapper[4676]: I0124 01:06:43.750672 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6d9458688d-h6hzt_ccb1ff12-bef7-4f23-b084-fae32f8202ac/manager/0.log" Jan 24 01:06:51 crc kubenswrapper[4676]: I0124 01:06:51.257120 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:06:51 crc kubenswrapper[4676]: E0124 01:06:51.259051 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:07:04 crc kubenswrapper[4676]: I0124 01:07:04.928209 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-mkwj8_f53afe6a-307c-4b0d-88cb-596703f35f8a/control-plane-machine-set-operator/0.log" Jan 24 01:07:05 crc kubenswrapper[4676]: I0124 01:07:05.086321 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jmv6d_9c51c973-c370-41e8-b167-25d3b11418bf/kube-rbac-proxy/0.log" Jan 24 01:07:05 crc kubenswrapper[4676]: I0124 01:07:05.125348 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-jmv6d_9c51c973-c370-41e8-b167-25d3b11418bf/machine-api-operator/0.log" Jan 24 01:07:06 crc kubenswrapper[4676]: I0124 01:07:06.271549 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:07:06 crc kubenswrapper[4676]: E0124 01:07:06.272020 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:07:18 crc kubenswrapper[4676]: I0124 01:07:18.805019 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-jws6v_1b51f181-5a85-4c07-b259-f67d17bf1134/cert-manager-controller/0.log" Jan 24 01:07:18 crc kubenswrapper[4676]: I0124 01:07:18.921410 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-s27fj_6db681a1-6165-4404-81a9-a189e9b30bfd/cert-manager-cainjector/0.log" Jan 24 01:07:19 crc kubenswrapper[4676]: I0124 01:07:19.002956 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-w7kb7_8f43788b-d559-4623-ae87-a820a2f23b08/cert-manager-webhook/0.log" Jan 24 01:07:21 crc kubenswrapper[4676]: I0124 01:07:21.256860 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:07:21 crc kubenswrapper[4676]: E0124 01:07:21.257497 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:07:32 crc kubenswrapper[4676]: I0124 01:07:32.345604 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-95988_4db34b09-85e6-435d-b991-c2513eec5d17/nmstate-console-plugin/0.log" Jan 24 01:07:32 crc kubenswrapper[4676]: I0124 01:07:32.494776 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lkqsz_6bdbc24e-ee52-41f0-9aaf-1091ac803c27/nmstate-handler/0.log" Jan 24 01:07:32 crc kubenswrapper[4676]: I0124 01:07:32.570717 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-rv2q7_0c99caac-d9eb-494c-bd04-c18dbc8a0844/kube-rbac-proxy/0.log" Jan 24 01:07:32 crc kubenswrapper[4676]: I0124 01:07:32.621553 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-rv2q7_0c99caac-d9eb-494c-bd04-c18dbc8a0844/nmstate-metrics/0.log" Jan 24 01:07:32 crc kubenswrapper[4676]: I0124 01:07:32.809412 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-jsbxx_efdaf1c0-0096-4f21-a0e2-2fc6f6e04de2/nmstate-operator/0.log" Jan 24 01:07:32 crc kubenswrapper[4676]: I0124 01:07:32.852051 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-cpbrv_badb3470-b60b-44ca-8d9e-52191ea016fa/nmstate-webhook/0.log" Jan 24 01:07:35 crc kubenswrapper[4676]: I0124 01:07:35.255749 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:07:35 crc kubenswrapper[4676]: E0124 01:07:35.256308 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:07:50 crc kubenswrapper[4676]: I0124 01:07:50.255616 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:07:50 crc kubenswrapper[4676]: E0124 01:07:50.256313 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.217033 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-jqk4f_f5517856-fac7-4312-ab46-86bbd5c1282d/controller/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.245003 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-jqk4f_f5517856-fac7-4312-ab46-86bbd5c1282d/kube-rbac-proxy/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.434782 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-frr-files/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.694780 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-reloader/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.727157 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-reloader/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.744151 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-metrics/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.752408 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-frr-files/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.883271 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-frr-files/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.888764 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-reloader/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.944922 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-metrics/0.log" Jan 24 01:08:00 crc kubenswrapper[4676]: I0124 01:08:00.987146 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-metrics/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.150359 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-frr-files/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.159359 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-metrics/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.187565 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/cp-reloader/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.191870 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/controller/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.442057 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/frr-metrics/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.450736 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/kube-rbac-proxy/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.475534 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/kube-rbac-proxy-frr/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.682050 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/reloader/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.758813 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-s7b59_2ce22d83-ee4f-4ad7-8882-b876d4ed52a2/frr-k8s-webhook-server/0.log" Jan 24 01:08:01 crc kubenswrapper[4676]: I0124 01:08:01.971739 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6868db74d6-qfhwb_59200cad-bd1a-472a-a1a1-adccb5211b21/manager/0.log" Jan 24 01:08:02 crc kubenswrapper[4676]: I0124 01:08:02.213181 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6dbffff8f5-cb5l5_adfe0dac-5ac5-44b8-97db-088d1ac83d34/webhook-server/0.log" Jan 24 01:08:02 crc kubenswrapper[4676]: I0124 01:08:02.285358 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zh6td_7da3019e-01de-4671-a78f-6c0d2e57fde3/kube-rbac-proxy/0.log" Jan 24 01:08:02 crc kubenswrapper[4676]: I0124 01:08:02.390238 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-szpqt_8a7798b0-f97f-4804-a406-22ecc8b45677/frr/0.log" Jan 24 01:08:02 crc kubenswrapper[4676]: I0124 01:08:02.681587 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zh6td_7da3019e-01de-4671-a78f-6c0d2e57fde3/speaker/0.log" Jan 24 01:08:04 crc kubenswrapper[4676]: I0124 01:08:04.256780 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:08:04 crc kubenswrapper[4676]: E0124 01:08:04.257801 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:08:16 crc kubenswrapper[4676]: I0124 01:08:16.267878 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:08:16 crc kubenswrapper[4676]: E0124 01:08:16.269022 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:08:18 crc kubenswrapper[4676]: I0124 01:08:18.481399 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/util/0.log" Jan 24 01:08:18 crc kubenswrapper[4676]: I0124 01:08:18.679144 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/pull/0.log" Jan 24 01:08:18 crc kubenswrapper[4676]: I0124 01:08:18.743853 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/pull/0.log" Jan 24 01:08:18 crc kubenswrapper[4676]: I0124 01:08:18.756827 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/util/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.017993 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/pull/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.029732 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/util/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.066117 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcrxg2j_3c111688-154b-47fa-8f89-6e48007b1fec/extract/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.217827 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/util/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.419178 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/pull/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.449015 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/pull/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.465647 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/util/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.656737 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/pull/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.748320 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/util/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.787231 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec713rcsfv_79811b63-a3e6-47b0-8041-247b1536ba50/extract/0.log" Jan 24 01:08:19 crc kubenswrapper[4676]: I0124 01:08:19.944237 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-utilities/0.log" Jan 24 01:08:20 crc kubenswrapper[4676]: I0124 01:08:20.275563 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-content/0.log" Jan 24 01:08:20 crc kubenswrapper[4676]: I0124 01:08:20.322243 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-utilities/0.log" Jan 24 01:08:20 crc kubenswrapper[4676]: I0124 01:08:20.339887 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-content/0.log" Jan 24 01:08:20 crc kubenswrapper[4676]: I0124 01:08:20.637922 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-utilities/0.log" Jan 24 01:08:20 crc kubenswrapper[4676]: I0124 01:08:20.684267 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/extract-content/0.log" Jan 24 01:08:20 crc kubenswrapper[4676]: I0124 01:08:20.812590 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-vsfsj_34862c1d-2d18-42f8-9ef7-71d349c019fd/registry-server/0.log" Jan 24 01:08:20 crc kubenswrapper[4676]: I0124 01:08:20.912837 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-utilities/0.log" Jan 24 01:08:21 crc kubenswrapper[4676]: I0124 01:08:21.397080 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-utilities/0.log" Jan 24 01:08:21 crc kubenswrapper[4676]: I0124 01:08:21.415527 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-content/0.log" Jan 24 01:08:21 crc kubenswrapper[4676]: I0124 01:08:21.458490 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-content/0.log" Jan 24 01:08:21 crc kubenswrapper[4676]: I0124 01:08:21.656299 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-utilities/0.log" Jan 24 01:08:21 crc kubenswrapper[4676]: I0124 01:08:21.664048 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/extract-content/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.058285 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-utilities/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.092017 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjjnx_646456dc-35bc-4df2-8f92-55cdfefc6010/registry-server/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.113682 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-kwptt_8db31db6-2c7c-4688-89c7-328024cd7003/marketplace-operator/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.354013 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-utilities/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.385593 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-content/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.429210 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-content/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.546034 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-utilities/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.689704 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/extract-content/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.756467 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-vw7tg_4c82da49-780b-431e-bfe7-d52ce3bcb623/registry-server/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.859866 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-utilities/0.log" Jan 24 01:08:22 crc kubenswrapper[4676]: I0124 01:08:22.993751 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-content/0.log" Jan 24 01:08:23 crc kubenswrapper[4676]: I0124 01:08:23.037506 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-utilities/0.log" Jan 24 01:08:23 crc kubenswrapper[4676]: I0124 01:08:23.072127 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-content/0.log" Jan 24 01:08:23 crc kubenswrapper[4676]: I0124 01:08:23.198126 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-content/0.log" Jan 24 01:08:23 crc kubenswrapper[4676]: I0124 01:08:23.243675 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/extract-utilities/0.log" Jan 24 01:08:23 crc kubenswrapper[4676]: I0124 01:08:23.624231 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6mz4x_829774a4-dd19-462e-829c-f201bddf6886/registry-server/0.log" Jan 24 01:08:27 crc kubenswrapper[4676]: I0124 01:08:27.255202 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:08:27 crc kubenswrapper[4676]: E0124 01:08:27.255971 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:08:39 crc kubenswrapper[4676]: I0124 01:08:39.255904 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:08:39 crc kubenswrapper[4676]: E0124 01:08:39.256750 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:08:53 crc kubenswrapper[4676]: I0124 01:08:53.255705 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:08:53 crc kubenswrapper[4676]: E0124 01:08:53.256621 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.274980 4676 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2nrdw"] Jan 24 01:08:58 crc kubenswrapper[4676]: E0124 01:08:58.275754 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerName="extract-content" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.275766 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerName="extract-content" Jan 24 01:08:58 crc kubenswrapper[4676]: E0124 01:08:58.275798 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerName="registry-server" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.275804 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerName="registry-server" Jan 24 01:08:58 crc kubenswrapper[4676]: E0124 01:08:58.275819 4676 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerName="extract-utilities" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.275825 4676 state_mem.go:107] "Deleted CPUSet assignment" podUID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerName="extract-utilities" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.275973 4676 memory_manager.go:354] "RemoveStaleState removing state" podUID="218ff773-95bb-4b63-a4d5-b7fb752b5870" containerName="registry-server" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.277141 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.297212 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2nrdw"] Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.374418 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf4b7\" (UniqueName: \"kubernetes.io/projected/3b5bdb57-1227-4549-a407-7768a82ff571-kube-api-access-gf4b7\") pod \"redhat-operators-2nrdw\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.374493 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-utilities\") pod \"redhat-operators-2nrdw\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.374513 4676 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-catalog-content\") pod \"redhat-operators-2nrdw\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.477213 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf4b7\" (UniqueName: \"kubernetes.io/projected/3b5bdb57-1227-4549-a407-7768a82ff571-kube-api-access-gf4b7\") pod \"redhat-operators-2nrdw\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.477521 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-utilities\") pod \"redhat-operators-2nrdw\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.477542 4676 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-catalog-content\") pod \"redhat-operators-2nrdw\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.478020 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-catalog-content\") pod \"redhat-operators-2nrdw\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.478080 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-utilities\") pod \"redhat-operators-2nrdw\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.499695 4676 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf4b7\" (UniqueName: \"kubernetes.io/projected/3b5bdb57-1227-4549-a407-7768a82ff571-kube-api-access-gf4b7\") pod \"redhat-operators-2nrdw\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:58 crc kubenswrapper[4676]: I0124 01:08:58.603183 4676 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:08:59 crc kubenswrapper[4676]: I0124 01:08:59.112592 4676 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2nrdw"] Jan 24 01:08:59 crc kubenswrapper[4676]: I0124 01:08:59.856541 4676 generic.go:334] "Generic (PLEG): container finished" podID="3b5bdb57-1227-4549-a407-7768a82ff571" containerID="4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b" exitCode=0 Jan 24 01:08:59 crc kubenswrapper[4676]: I0124 01:08:59.856650 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nrdw" event={"ID":"3b5bdb57-1227-4549-a407-7768a82ff571","Type":"ContainerDied","Data":"4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b"} Jan 24 01:08:59 crc kubenswrapper[4676]: I0124 01:08:59.857004 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nrdw" event={"ID":"3b5bdb57-1227-4549-a407-7768a82ff571","Type":"ContainerStarted","Data":"d2b9c41fdecc2fd86c961eaa1c90432f775d4cec810f3d698cf239294c2169aa"} Jan 24 01:09:00 crc kubenswrapper[4676]: I0124 01:09:00.871721 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nrdw" event={"ID":"3b5bdb57-1227-4549-a407-7768a82ff571","Type":"ContainerStarted","Data":"706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee"} Jan 24 01:09:04 crc kubenswrapper[4676]: I0124 01:09:04.903923 4676 generic.go:334] "Generic (PLEG): container finished" podID="3b5bdb57-1227-4549-a407-7768a82ff571" containerID="706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee" exitCode=0 Jan 24 01:09:04 crc kubenswrapper[4676]: I0124 01:09:04.904005 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nrdw" event={"ID":"3b5bdb57-1227-4549-a407-7768a82ff571","Type":"ContainerDied","Data":"706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee"} Jan 24 01:09:05 crc kubenswrapper[4676]: I0124 01:09:05.914297 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nrdw" event={"ID":"3b5bdb57-1227-4549-a407-7768a82ff571","Type":"ContainerStarted","Data":"93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e"} Jan 24 01:09:05 crc kubenswrapper[4676]: I0124 01:09:05.937226 4676 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2nrdw" podStartSLOduration=2.355179058 podStartE2EDuration="7.93720857s" podCreationTimestamp="2026-01-24 01:08:58 +0000 UTC" firstStartedPulling="2026-01-24 01:08:59.859035085 +0000 UTC m=+3923.889006086" lastFinishedPulling="2026-01-24 01:09:05.441064597 +0000 UTC m=+3929.471035598" observedRunningTime="2026-01-24 01:09:05.930787054 +0000 UTC m=+3929.960758095" watchObservedRunningTime="2026-01-24 01:09:05.93720857 +0000 UTC m=+3929.967179571" Jan 24 01:09:06 crc kubenswrapper[4676]: I0124 01:09:06.279791 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:09:06 crc kubenswrapper[4676]: E0124 01:09:06.280462 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:09:08 crc kubenswrapper[4676]: I0124 01:09:08.603303 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:09:08 crc kubenswrapper[4676]: I0124 01:09:08.603563 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:09:09 crc kubenswrapper[4676]: I0124 01:09:09.672277 4676 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2nrdw" podUID="3b5bdb57-1227-4549-a407-7768a82ff571" containerName="registry-server" probeResult="failure" output=< Jan 24 01:09:09 crc kubenswrapper[4676]: timeout: failed to connect service ":50051" within 1s Jan 24 01:09:09 crc kubenswrapper[4676]: > Jan 24 01:09:17 crc kubenswrapper[4676]: I0124 01:09:17.255811 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:09:17 crc kubenswrapper[4676]: E0124 01:09:17.257072 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:09:18 crc kubenswrapper[4676]: I0124 01:09:18.694161 4676 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:09:18 crc kubenswrapper[4676]: I0124 01:09:18.747463 4676 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:09:18 crc kubenswrapper[4676]: I0124 01:09:18.939604 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2nrdw"] Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.043914 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2nrdw" podUID="3b5bdb57-1227-4549-a407-7768a82ff571" containerName="registry-server" containerID="cri-o://93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e" gracePeriod=2 Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.549787 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.613047 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf4b7\" (UniqueName: \"kubernetes.io/projected/3b5bdb57-1227-4549-a407-7768a82ff571-kube-api-access-gf4b7\") pod \"3b5bdb57-1227-4549-a407-7768a82ff571\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.613090 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-catalog-content\") pod \"3b5bdb57-1227-4549-a407-7768a82ff571\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.613217 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-utilities\") pod \"3b5bdb57-1227-4549-a407-7768a82ff571\" (UID: \"3b5bdb57-1227-4549-a407-7768a82ff571\") " Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.614407 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-utilities" (OuterVolumeSpecName: "utilities") pod "3b5bdb57-1227-4549-a407-7768a82ff571" (UID: "3b5bdb57-1227-4549-a407-7768a82ff571"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.614663 4676 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-utilities\") on node \"crc\" DevicePath \"\"" Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.627269 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b5bdb57-1227-4549-a407-7768a82ff571-kube-api-access-gf4b7" (OuterVolumeSpecName: "kube-api-access-gf4b7") pod "3b5bdb57-1227-4549-a407-7768a82ff571" (UID: "3b5bdb57-1227-4549-a407-7768a82ff571"). InnerVolumeSpecName "kube-api-access-gf4b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.717003 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf4b7\" (UniqueName: \"kubernetes.io/projected/3b5bdb57-1227-4549-a407-7768a82ff571-kube-api-access-gf4b7\") on node \"crc\" DevicePath \"\"" Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.738388 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b5bdb57-1227-4549-a407-7768a82ff571" (UID: "3b5bdb57-1227-4549-a407-7768a82ff571"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:09:20 crc kubenswrapper[4676]: I0124 01:09:20.818554 4676 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b5bdb57-1227-4549-a407-7768a82ff571-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.058684 4676 generic.go:334] "Generic (PLEG): container finished" podID="3b5bdb57-1227-4549-a407-7768a82ff571" containerID="93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e" exitCode=0 Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.058766 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nrdw" event={"ID":"3b5bdb57-1227-4549-a407-7768a82ff571","Type":"ContainerDied","Data":"93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e"} Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.058791 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2nrdw" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.058809 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2nrdw" event={"ID":"3b5bdb57-1227-4549-a407-7768a82ff571","Type":"ContainerDied","Data":"d2b9c41fdecc2fd86c961eaa1c90432f775d4cec810f3d698cf239294c2169aa"} Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.058844 4676 scope.go:117] "RemoveContainer" containerID="93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.090214 4676 scope.go:117] "RemoveContainer" containerID="706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.125071 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2nrdw"] Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.125897 4676 scope.go:117] "RemoveContainer" containerID="4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.140129 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2nrdw"] Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.164441 4676 scope.go:117] "RemoveContainer" containerID="93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e" Jan 24 01:09:21 crc kubenswrapper[4676]: E0124 01:09:21.164948 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e\": container with ID starting with 93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e not found: ID does not exist" containerID="93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.164998 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e"} err="failed to get container status \"93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e\": rpc error: code = NotFound desc = could not find container \"93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e\": container with ID starting with 93c3dc5d89006d93c816256ff7e52391b95edbe86e4b6bff514031d0a377503e not found: ID does not exist" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.165031 4676 scope.go:117] "RemoveContainer" containerID="706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee" Jan 24 01:09:21 crc kubenswrapper[4676]: E0124 01:09:21.165821 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee\": container with ID starting with 706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee not found: ID does not exist" containerID="706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.165952 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee"} err="failed to get container status \"706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee\": rpc error: code = NotFound desc = could not find container \"706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee\": container with ID starting with 706f2526bdcdb0475f55e82716fff7b99c6f21ba4ce5453d4af3afdd341ddbee not found: ID does not exist" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.166058 4676 scope.go:117] "RemoveContainer" containerID="4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b" Jan 24 01:09:21 crc kubenswrapper[4676]: E0124 01:09:21.166470 4676 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b\": container with ID starting with 4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b not found: ID does not exist" containerID="4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b" Jan 24 01:09:21 crc kubenswrapper[4676]: I0124 01:09:21.166508 4676 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b"} err="failed to get container status \"4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b\": rpc error: code = NotFound desc = could not find container \"4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b\": container with ID starting with 4cbfe82eff6d462ba3148ce02c0da1c0abc3bda8a2446df0a50c15b4fa35d06b not found: ID does not exist" Jan 24 01:09:22 crc kubenswrapper[4676]: I0124 01:09:22.267196 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b5bdb57-1227-4549-a407-7768a82ff571" path="/var/lib/kubelet/pods/3b5bdb57-1227-4549-a407-7768a82ff571/volumes" Jan 24 01:09:29 crc kubenswrapper[4676]: I0124 01:09:29.259120 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:09:29 crc kubenswrapper[4676]: E0124 01:09:29.259788 4676 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7mzrz_openshift-machine-config-operator(bd647b0d-6d3d-432d-81ac-6484a2948211)\"" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" Jan 24 01:09:40 crc kubenswrapper[4676]: I0124 01:09:40.257728 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:09:41 crc kubenswrapper[4676]: I0124 01:09:41.340868 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"edd2934dbaa930bc821862a558e8c2c3faf93460442893d12bfeffbb6ed69162"} Jan 24 01:10:18 crc kubenswrapper[4676]: I0124 01:10:18.784938 4676 generic.go:334] "Generic (PLEG): container finished" podID="ce7f9d18-7eec-43f4-908a-d65d210492e7" containerID="d0b712dc713fcee1a9dcfc5340b4f493d3bd6dca5b50bbf686427d3022f3f332" exitCode=0 Jan 24 01:10:18 crc kubenswrapper[4676]: I0124 01:10:18.785058 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pgb89/must-gather-rzdhh" event={"ID":"ce7f9d18-7eec-43f4-908a-d65d210492e7","Type":"ContainerDied","Data":"d0b712dc713fcee1a9dcfc5340b4f493d3bd6dca5b50bbf686427d3022f3f332"} Jan 24 01:10:18 crc kubenswrapper[4676]: I0124 01:10:18.786489 4676 scope.go:117] "RemoveContainer" containerID="d0b712dc713fcee1a9dcfc5340b4f493d3bd6dca5b50bbf686427d3022f3f332" Jan 24 01:10:19 crc kubenswrapper[4676]: I0124 01:10:19.171841 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pgb89_must-gather-rzdhh_ce7f9d18-7eec-43f4-908a-d65d210492e7/gather/0.log" Jan 24 01:10:30 crc kubenswrapper[4676]: I0124 01:10:30.755767 4676 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pgb89/must-gather-rzdhh"] Jan 24 01:10:30 crc kubenswrapper[4676]: I0124 01:10:30.756606 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pgb89/must-gather-rzdhh" podUID="ce7f9d18-7eec-43f4-908a-d65d210492e7" containerName="copy" containerID="cri-o://ebdf94432be40adc9409ec2d7d078306d66cbfc22f6dc35636a6fecb28c07f9e" gracePeriod=2 Jan 24 01:10:30 crc kubenswrapper[4676]: I0124 01:10:30.769066 4676 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pgb89/must-gather-rzdhh"] Jan 24 01:10:30 crc kubenswrapper[4676]: I0124 01:10:30.905741 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pgb89_must-gather-rzdhh_ce7f9d18-7eec-43f4-908a-d65d210492e7/copy/0.log" Jan 24 01:10:30 crc kubenswrapper[4676]: I0124 01:10:30.906442 4676 generic.go:334] "Generic (PLEG): container finished" podID="ce7f9d18-7eec-43f4-908a-d65d210492e7" containerID="ebdf94432be40adc9409ec2d7d078306d66cbfc22f6dc35636a6fecb28c07f9e" exitCode=143 Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.256177 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pgb89_must-gather-rzdhh_ce7f9d18-7eec-43f4-908a-d65d210492e7/copy/0.log" Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.256958 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.405259 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/ce7f9d18-7eec-43f4-908a-d65d210492e7-kube-api-access-z2kcb\") pod \"ce7f9d18-7eec-43f4-908a-d65d210492e7\" (UID: \"ce7f9d18-7eec-43f4-908a-d65d210492e7\") " Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.405497 4676 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ce7f9d18-7eec-43f4-908a-d65d210492e7-must-gather-output\") pod \"ce7f9d18-7eec-43f4-908a-d65d210492e7\" (UID: \"ce7f9d18-7eec-43f4-908a-d65d210492e7\") " Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.411555 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce7f9d18-7eec-43f4-908a-d65d210492e7-kube-api-access-z2kcb" (OuterVolumeSpecName: "kube-api-access-z2kcb") pod "ce7f9d18-7eec-43f4-908a-d65d210492e7" (UID: "ce7f9d18-7eec-43f4-908a-d65d210492e7"). InnerVolumeSpecName "kube-api-access-z2kcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.507958 4676 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2kcb\" (UniqueName: \"kubernetes.io/projected/ce7f9d18-7eec-43f4-908a-d65d210492e7-kube-api-access-z2kcb\") on node \"crc\" DevicePath \"\"" Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.540330 4676 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce7f9d18-7eec-43f4-908a-d65d210492e7-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "ce7f9d18-7eec-43f4-908a-d65d210492e7" (UID: "ce7f9d18-7eec-43f4-908a-d65d210492e7"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.609359 4676 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/ce7f9d18-7eec-43f4-908a-d65d210492e7-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.916017 4676 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pgb89_must-gather-rzdhh_ce7f9d18-7eec-43f4-908a-d65d210492e7/copy/0.log" Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.916604 4676 scope.go:117] "RemoveContainer" containerID="ebdf94432be40adc9409ec2d7d078306d66cbfc22f6dc35636a6fecb28c07f9e" Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.916719 4676 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pgb89/must-gather-rzdhh" Jan 24 01:10:31 crc kubenswrapper[4676]: I0124 01:10:31.940723 4676 scope.go:117] "RemoveContainer" containerID="d0b712dc713fcee1a9dcfc5340b4f493d3bd6dca5b50bbf686427d3022f3f332" Jan 24 01:10:32 crc kubenswrapper[4676]: I0124 01:10:32.277616 4676 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce7f9d18-7eec-43f4-908a-d65d210492e7" path="/var/lib/kubelet/pods/ce7f9d18-7eec-43f4-908a-d65d210492e7/volumes" Jan 24 01:11:43 crc kubenswrapper[4676]: I0124 01:11:43.845657 4676 scope.go:117] "RemoveContainer" containerID="6e1d712a5b60b5fed2dd683d40f7c0b75205c48776013f310795ca214bdf4a77" Jan 24 01:12:09 crc kubenswrapper[4676]: I0124 01:12:09.363927 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 01:12:09 crc kubenswrapper[4676]: I0124 01:12:09.364636 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 01:12:39 crc kubenswrapper[4676]: I0124 01:12:39.364779 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 01:12:39 crc kubenswrapper[4676]: I0124 01:12:39.365321 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 01:13:09 crc kubenswrapper[4676]: I0124 01:13:09.364313 4676 patch_prober.go:28] interesting pod/machine-config-daemon-7mzrz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 24 01:13:09 crc kubenswrapper[4676]: I0124 01:13:09.365055 4676 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 24 01:13:09 crc kubenswrapper[4676]: I0124 01:13:09.365131 4676 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" Jan 24 01:13:09 crc kubenswrapper[4676]: I0124 01:13:09.366068 4676 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"edd2934dbaa930bc821862a558e8c2c3faf93460442893d12bfeffbb6ed69162"} pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 24 01:13:09 crc kubenswrapper[4676]: I0124 01:13:09.366134 4676 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" podUID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerName="machine-config-daemon" containerID="cri-o://edd2934dbaa930bc821862a558e8c2c3faf93460442893d12bfeffbb6ed69162" gracePeriod=600 Jan 24 01:13:09 crc kubenswrapper[4676]: I0124 01:13:09.626998 4676 generic.go:334] "Generic (PLEG): container finished" podID="bd647b0d-6d3d-432d-81ac-6484a2948211" containerID="edd2934dbaa930bc821862a558e8c2c3faf93460442893d12bfeffbb6ed69162" exitCode=0 Jan 24 01:13:09 crc kubenswrapper[4676]: I0124 01:13:09.627040 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerDied","Data":"edd2934dbaa930bc821862a558e8c2c3faf93460442893d12bfeffbb6ed69162"} Jan 24 01:13:09 crc kubenswrapper[4676]: I0124 01:13:09.627073 4676 scope.go:117] "RemoveContainer" containerID="6b709e3553e881b387af6667b2735e4c905d3b5ba75799e955a72575129b29d9" Jan 24 01:13:10 crc kubenswrapper[4676]: I0124 01:13:10.643418 4676 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7mzrz" event={"ID":"bd647b0d-6d3d-432d-81ac-6484a2948211","Type":"ContainerStarted","Data":"64412b3d11bc4a5ba20e991951bf85349b15c728095f990ca935e7e9fc752a77"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515135016331024443 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015135016332017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015135005530016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015135005530015452 5ustar corecore